var/home/core/zuul-output/0000755000175000017500000000000015150342034014522 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015150353633015475 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000323464215150353551020267 0ustar corecoreiסikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs,r.k9Gf܅A"mv?_eGbuuțx{w7ݭ7֫\% oo/q3m^]/o?8.7oW}ʋghewx/mX,ojŻ ^Tb3b#׳:}=p7뼝ca㑔`e0I1Q!&ѱ[/o^{W-{t3_U|6 x)K#/5ΌR"ggóisR)N %emOQ/Ϋ[oa0vsT*v (6pk**+ Le*gUWi [ӊg*XCF*A(-aD~JwFPO7M$n6iXύO^%26lDt#3{f!f6;WR.!$5 J:1*S%V!F([EbD]娍ԹiE03`Cfw&:ɴ@=yN{f}\{+>2^G) u.`l(Sm&F4a0>eBmFR5]!PI6f٘"y/(":[#;`1}+7 s'ϨF&%8'# $9b"r>B)GF%\bi/ Ff/Bp 4YH~BŊ6EZ|^߸3%L[EC 7gg/碓@e=Vn)h\\lwCzDiQJxTsL] ,=M`nͷ~Vܯ5n|X*ǘ;RJK!b>JR*kl|+"N'C_#a7]d]sJg;;>Yp׫,w`ɚ'd$ecwŻ^~7EpQС3DCS[Yʧ?DDS aw߾)VxX帟AB}nyи0stĈCo.:wAZ{sy:7qsWctx{}n-+ZYsI{/.Ra9XcђQ0FK@aEDO2es ׇN# ZF͹b,*YVi+$<QMGhC}^}?BqG!(8l K3T[<~6]90}(*T7siv'=k 9Q2@vN ( R['>v*;o57sp$3ncx!>t®W>]tF-iܪ%GYbaRvHa}dkD̶*';ک|s_}8yj,('GrgTZ'U鋊TqOſ * /Ijo!՟8`"j}zӲ$k3jS|C7;A)͎V.r?t\WU1ojjr<~Tq> `=tJ!aݡ=h6Yݭw}?lѹ`f_" J9w4ts7NG GGG]ҡgc⌝M b/Ζdžqߎ5.)Q VE JN`:a!KM/+9 bG+މG uIo1]ߔr TGGJ\B BR 4X\r RYGVق?<6jHSJ Jno#ˏl_}z?1:N3cl.:f 3 JJ5Z|&הԟ,Tصp&NI%`t3Vi=Ob㸵2*3d*mQ%"h+ "f "D(~~moH|E3*46$Ag4aX)Ǜƾ9U Ӆ^};ڲ7J9@ kV%g>a~W;D=;y|AAY'"葋_d$Ə{(he NSfX1982TH#D֪v3l"<, { Tms'oI&'Adp]{1DL^5"Ϧޙ`F}W5XDV7V5EE9esYYfiMOV i/ f>3VQ 7,oTW⇊AqO:rƭĘ DuZ^ To3dEN/} fI+?|Uz5SUZa{P,97óI,Q{eNFV+(hʺb ״ʻʞX6ýcsT z`q 0C?41- _n^ylSO2|'W'BOTLl-9Ja [$3BV2DC4l!TO C*Mrii1f5 JA *#jv߿Imy%u LOL8c3ilLJ!Ip,2(( *%KGj   %*e5-wFp"a~fzqu6tY,d,`!qIv꜒"T[1!I!NwL}\|}.b3oXR\(L _nJB/_xY.# ſԸv}9U}'/o uSH<:˷tGLS0l/LKcQ.os2% t)Eh~2p cL1%'4-1a_`[Zz㧦|k˭c ĚOρ_} Ewt3th?tvͪ{~;J0= |JUԍ;Iw}/9nh7l%>'ct Հ}a>-:(QxPyA Z UcÖgڌ:8cΗ|U1,-N9 dI [@3YNV ̍"ޛ4tO,{=hFѓ$b =D(zn;Y<1x~SJ^{vn 9 j1шk'L"cE=K]A(oQ۲6+ktwLzG,87^ 9H\yqū1)\(v8pHA"ΈGVp"c ?Z)hm.2;sl$瓴ӘIe~H|.Y#C^SJĽHǀeTwvy"v܅ ]?22R.lQPa ˆSܫ1z.x62%z].`Gn&*7bd+, Z`ͲH-nမ^WbPFtOfD]c9\w+ea~~{;Vm >|WAޭi`HbIãE{%&4]Iw Wjoru ݜmKnZ<X; ۢ( nx K8.|DXb +*598;w)zp:̊~;͞)6vnM!N5Cu!8Wq/`FUwWAֻ,Qu W@ Fi:K [Av*_958]a:pmQ&'ᚡmi@ zF(n&P;)_]µ!doR0`pl`~9Fk[ٺ+4Hhao-jϸ??R<lb#P-^39T|L /~p│x@Bq"M/lja\b݋af LnU*P(8W[U6WX ZoѶ^SH:K:%Qvl\b FqQI.ȨHWo;Nw$͹O$oEE-eq=.*Dp,V;(bgJ!gF)892sw*+{[or@x,))[o新#.͞.;=fc<)((b۲Eumw峛M2,V[cm,S~ AF~.2v?JNt=O7^r.@DEuU1}g$>8ac#sĢB\PIPfwJQJ;Qxm &GBf\ZA$Ba-z|A-I @x70 晪MV)m8[6-Te@`E|=U D(C{oVa*H7MQK"<O%MTTtx袥:2JޚݶKd7UZihRk71VDqiގ\<:Ѓ3"gJJčE&>&EI|I˿j2ǯɘCGOa9C1L ={fm&'^tigk$DA' elW@Tiv{ !]oBLKJO*t*\n-iȚ4`{x_z;j3Xh ׄ?xt.o:`x^d~0u$ v48 0_ | E"Hd"H`A0&dY3 ً[fctWF_hdxMUY.b=eaI3Z=᢬-'~DWc;j FRrI5%N/K;Dk rCbm7чsSW_8g{RY.~XfEߪg:smBi1 YBX4),[c^54Sg(s$sN' 88`wC3TE+A\.ԍל9 y{͝BxG&JS meT;{З>'[LR"w F05N<&AJ3DA0ʄ4(zTUWDdE3̻l^-Xw3Fɀ{B-~.h+U8 i1b8wؖ#~zQ`/L 9#Pu/<4A L<KL U(Ee'sCcq !Ȥ4΍ +aM(VldX ][T !Ȱ|HN~6y,⒊)$e{)SR#kהyϛ7^i58f4PmB8 Y{qeφvk73:1@ƛ.{f8IGv*1藺yx27M=>+VnG;\<x7v21՚H :[Γd!E'a4n?k[A׈(sob 41Y9(^SE@7`KIK`kx& V`X0,%pe_ן >hd xе"Q4SUwy x<'o_~#6$g!D$c=5ۄX[ു RzG:柺[ӏ[3frl ô ހ^2TӘUAT!94[[m۾\T)W> lv+ H\FpG)ۏjk_c51̃^cn ba-X/#=Im41NLu\9ETp^poAOO&AtswMmeüze1NHr~53݃u$J!̖G8>pzW%6TXZZlRcI:fsU k[*'c}eΫ g콬)W{_k R0gZ4A%%AA.! J${4od9`mCX:Q|ɠ(" H-ksTQFF]gwX :]̼o}%D*!`HhK&TQ5<peS0Q{nX!S1̒t&LJ&YE)0]@ 2φ ngd, =S;Z&y ^]\=5Ϸe4>#HPU5j6-MSv`F|Ʈ;l:!3磃F8>q|ho4UXk.];DW$%kWdud/ Fق%EBp$CY#$R^thYxKG79*yäL0 ;_4<7U\戌g=R:y #Mlc!MH dMʳ!ٞeˮt6Z֙Fq.w/0vŵlI=V.I=^nէ賔‡,T4U˛y$%lrK2f1X+x焳!61Bǝ8f><ON(k2H`P k ݊14]L*SHh瀷[\Iuu#xW~N/H^wFIm)1 X{{\@Kɩx3']x'/AGl#UQk*et1v0_YG@6N{G/ _@a>GkYtUXeBIn8P2PN ԓ^]cc{c2:P逵Q4w@)k{ {ZႷٱ6HY;ޣP(uY$SAك DڳI%CLȺ($j#}AYѪ)'z ftqLd#,#c;!Sp+.;s7$ G5]n2~SKnv@P@}X;ó*KY;<˥C*&gK×>| @+`^weR.bd]r߈;J9g8,?tPXUYDP` ze:g h4Yqst۷ خ!s5s "ќxO6h2K2w6l#u|Q)ɝ.u(2(Іu=\]<9Ӛ5mݢJuKE_4pσw]˳ϛA+eݥƃJyO 2&XqIr=%R2wË- z"Uw_ ޭK&وrNb}SȘNHM9VVY{ 2@@gx@6N% 2 9zI*(ƢIz'LjxzU_Fۙݨ*N84YܮU+"/L=fBL8)QRNw7.ƽnr;KO oYalm_hJڑ(ra>CFRKT|#Ҹ'؞ե=ugD$F?9%]4(@gth k67VVYWGh)OioFޞޙM{L+RC`aH^N.>ӻA| GG<&\'ؘ 1AЋ4YۣxL<9r7'9SHwGR-t7&Wtc܅ 9I S?l:R3ct1n$D`TV`Y{-HY#U|zcku^dv4`0tZL7%cCڣȀ5Wc8|ۘ<$ޟPi $Ϻ1D q:$yۨ%udz;sumק$m^dCw_軚Zd)9L$~ŻifqrӮKB}ݶ8iysNnOQ:kE,%k,:-25c4bmI?-<0O^Ϻy%Y()&{ G9ME`!3@P%a~Čy7q4'f T'} 6gБx7>xuQm2^ެGfDR|e3`kڕ(@N[b i;i5-Q}ֺM|doB$޵d"XSB 2J oB!3 #mU6~WI|~NDڮr> ol9e_inJtY:q1Bֱ};qbU[g_TRy@:ΘȜj}jV7j#oSDk"Cfnk8XU a}1{?,9GzT})ĢY QȞ';i|7m qVF5%.l)Ɩҏjkٴ*fVg,2Cm .߷*bd .vޜv:8F+t|fb.!aIA`]t˫m_ß]vS͝2磪l׳1XU7.v3QĞ5ly(Nɏ" $M~?M] ߋ,$%aX kQza5Z΂ 1<= 6s.8z% IXߊ[+}$"4Vj.&Ӎ61/I7}fbnf:K*2=+a2UeK5cWh(m!= \IhR#5!@7κR+N7lY]Fj=/ VΉYl_4UifD6~oFXJOVz' :lvt sJMQ6i*e"< Y?54hw'R`R1XY sXiRJ9sȊ,yPVs'SM}/}66xAU7IȆ'lspt +|2{u^)63}5 1=>gR}gs9QL߇!ӦxvbHJHe^ni5(~p>Ob?vUWzOvnEg\pS Ɨ$yKQ)sn*&<}mV Ѣ9Y;wQ ۊ*y@:\/z!<}Kj#Z"DWOYoU>Ie'BA$K* _RU[9 yJtOcMʹ4BZdL#-@?olJ'U\iϣy'u(9 WO{} tC:_}!ɢ}5"58}^ZXSŋz@hc$Nmy!6)F$%E,1?Ae!O.pꠅ`+^^"Lh28*f6^l bMMwO+Q?Fe\gzr~N$z׃%}WD΃d-W]v'K}wlZ9,)xIJPѥ djBXʇ);_x9]%.i|RY0dMd1{g&ζ+Ҥrx ђ{L?z0US4yBc#_=cƶ52R.)L)pŰl+ a[s7 Iz Ǫzd;-s~K7M>%} -Of;M.~P 8'k01Ѥ1HIa6Pn{/2ΏL+ΆhBUx5|T!F[|өSfFH.İd/D!-Ɩ:;v8`vU~Il2;VI]|Lu>$X(6 b ?u==O!;(>hǖVa[|oiya+CTm>C9|H iHe"j.S֔(*Cj!);Sak*ep~K1 v']7/.7 !ې: %ƶ(f갱/p  |T!|ik3cWW/ @a#ӸvZ{Ibi/b;u8IRXAV{ύԦٖwŅjIL{3#iyy >Vc11*Y0\N*HƽŇKoA`d;ɯw"O-]J"ȜI*DۂgؗN^saͭ̍*tPM*9aJ_ 3IVnס|< aUd⧘pvzz0V fN:ǖ9dɹt^dnJna) H _KӆX#rrE#r?uQ { xRF(߯y? jO]5_C!l]>a55[c&-W`a}TQX&mw*Ǫn\7{ctm,c%jP˃m )lwۨKqu!*ottonY77ܩJ==\J=]?Ww?¯8nq~q?A-T_qOq?5-3 |q |w.dަ&/<_ DVi^9 hxh2 Iz b.E)͢Q l1:YɊ",8'`*>q/E :Xd,RLW"Id9JogT\1f3@KuJ&@B x,A k ޒd [Yj-Ah1T9!(*t 0gb@񺱥-kc6V'“5huՂUmpa.% qZBh]Q; 'd:|ؒ3$".meO>Y?HELkYZP=8YAc| w#Dr) "h l`2@K$`#NXtJ^ zDpC6-]K[r0Z;`^ˁ-G$\~%Q;e{/d ^ ޒg0uE~ۊ$q9`尻]T#CJ1Ǐ9?M8]o2seXVt=ev!`JU#y8B*kM0{'\ 2n[{!fRБBmLaKfKywdb񱍠z{(.>LC,HI~'./bKjoJdpH UDp.cj|>z '` |]}4:q!G`G qBPu(DihU9P!`NHɩ݉S-^pşCx$BBRoJ@ѥuȑz.#&UݠmF̤@U8$ M6MY0/r: *s5xgs͙$ԙy#Ejl1#XX۾;R;+[E&Xi>eIi5lݍ )`8dM-}\\%.}T@ iS*XK!?\+ xJpΕ`p~mΖK Wu7Pll{f_WJp)h9U|A܌%`; TnpR4޷V+vy]/ϧ]+GЕ5҇#t~T)=UFEnvD8cRөcp6Rcc6:$[|038F*0-)ZyT10:U[tp޴}{~Y(f 4[m6F"roe5$!;VfBs˞ޝ4cc1ۀs`*f'r[8ݝYvjҹJ+0v yg[~)5 [j+Ag"pZ:"ka]+n!e߭lɹ$k'9~ J>0E8bJDƖ|e=rv:0e7>& 2ovN21cEdA Od[=jlV#XJ|&+-T1m8NP٤KX)tr:mDWx"8B*4*X FQG>^6 vq!EwQű&؁64Ĥj9| Οڭ:kg wa`e[GX$"JX!8j0"| \56cdʰHdX?"}B= -/%!C`@ шv1\h):=m%랹m RD3Q{]pcfՅuБ:A ѹ @˿ޗ~7e3tj>Y)"\**vdP=I6p;bck[ RhT#N0d5+A>ΰ-8sѹ Ve掟^ CZQc~j\b8$4kJ^آX/ 2z .}'1"+L=$ÅjuƖ},X n*[hp9 n`g.  RG-m~\y[j_;3\弁^bD5p-^〩:w}[ą8dBմVsrAJsT=~#0t.P*2V q%so#r|.v\sfa:\X%;3Xl; ՈC.5Wg󵸊y!1U:pUC4Cmp-7t]斻38ѮIWί_#z7u&Ӄcx-w+LX)w>^ʮٹUg:lR@djӓab u[kWw{7st28bJ0U1|z:9lX)zS&QsTomDvU`tiz5Ӄ~ 5yx `iݗE@Ubc@ ۾S6p{dMVwfa}/TQXȴ7Ij.WU}-I ux1^_VgϽeՠyq9Elq.EB*֕Wa 4GWO cUE7*1>b,|_›q< Q2=?10͓jB/ `8 vz2~SR퀧b@ !2t {i|r[/SSŕi!Ba;?7W'oxp\^W4 nx`mh2\i`y`y`Qr?F&!øQ\{hyVT3Gsv^6?/tAw%W>|Y8?J:`0'=ZﱟZ9Lx/Hrrr.u|6qCAxA:y?# L KOoj/Eu] %C| 4/I_Ѓ@zxpzbF`h|95no[#yDaHx>[1MڗZ ^ Mϻ!9wsp>NNK\v[{I) _yT4W~QZP0Eu0ϛG?#eS@k#˵uX}ON,۴MK$_ ݗר[p,"_h>!ˏ \Þ.n;Fe2w8 ?;yU3%-QXͬu%ŽC z*/Fs^-|1-3p1 K퀡5ﬢͣ"kg>3Bfǡ`oϡ|7 ̴&7Y ;t\Q3?[ę$3 ܮK4|=)MwGI\B ZluYę ȴ.x,y4|0Kq#`9 ;bo1_L4m˚VFDYyB@DjeTNr^ˬA|b"eZI\šRl7_,\ɲ_*^^)-<*6{KI$Je.;L-I+)UUu%G)'aa&f2 /x0ű`!lb%䓰j/q/m/,zZuݓ_r v|Gϣ/¹~Ba K15?'e5lX4IKhk O?=.ltn3J #?W51'GO)A 2aDCXSވei UGϻns](Uqq\\Ukb<ϗoIYG.BDGA\|M ޞ0PLNr΋__'AI񧬖suPQLw"ys08Xc ?v>lco@ʴkC Xu漎h^{#쿼2Z,Ssk/f kLN  * M\G t8ilgi" _pp?|5 ?ܭ\1Φ`m7YV5vĮ <1NMkC`#:+ރ'9- RoyABV9haIyMbu߶}E:GY0~]A98,՟IoCONy\~q!W/r7wHp6f6*o Wӥn546r.`q;> #$9goOFqi3xa?2:0o IJ{ۈ/Ҭ"{ezr }uvQ_\,@0H:o6f;(@[sRWY"ӇmQPVqU!xrȶ Y p`73^tfM-0GAAЙAM'ɳV4.3t;Upa%_ZDa:l,NݜV0,EW[wP3D ^W"{wv^Ňo #Q0qqo2 _<_<ʯ6]Le` Y@ 8ؾEAa8;nc{<qKbqє]u7X<2[$7z䋥ȯjܖy rQPDN<N+C̚ ]F紭Lɺ@B2d8 VrYلň܍bؔ$='ƃ@:GE j0\@ӒBU G״t,]'Y]j*p-GA1㮢GKZ%|yqE$ ) r@TٕHk-(qn]sh̀Y|DD!tF]v]K@|x8ۓ]7hDaEAh>Cp,(hjfy3 +Y Y2ߺFP ~&c[#Dn?á1iQmuh[* Y!M`u|Ծ `^D x G8RCh_rrG` eVӑ\5 hz/ZycZ8jAl$qX֨GyqX/Q2M%Ryʬp|rzS~ӬSfs閰Jgm7ov[= yYrYO@ho+Ӿ@b2><+G0Q6zP3xfyFbW!15?,/mС!(7( (A8iw6#sjYszZ"-c+>ص#m+-Ri5`C cUs^Ջ,(mebq^e 7Yvdf/b&#zKl9j.Qeb%k.+.EK.$)`k.{quoo_ R)nS$1ɛڋDоԻv`c%Xݾ"e&.V<-wlfz_og!v1(ՎpUy -*iKxJk" 0`giz:I<0SB]he ªX^\DWEݑ&J8U 5_^ B|E*}j5p-WBҳx N\3% ڝ&@D_#& Hq\G4=3ݜ8'qY--AUq>B5T瓲[59! #2SK U`bT}ְdN6-H捥"N -'eXj2pLaI8ήE9dKw6ň2񡖈_@LZmk aEꌵhJ{IZ`vjf42qX_\AmӔ ab7inO s0MDD=DEL jxn3>ϹUk-*[hwKM?]]ULKx{EUT \&&Qz^2 anpvZVO%Gݒ݃r%nyv̞Utuo+r(Lu~u;eEpW[xu*+ 7 R+ 20bQ(JiŘ*˴KKƊ˜<J! 3-oO#j`X9IVdxg@YåcE)ھP=!bՙ1VO7>>).Sm1]M]U%FT.iA%_ʺxQdߞfkly걯%\VOg'Yyϐ2d})+h$>SҭWb׵fLU; 8<+ϙ#GG}ѭi=YڊMu˲brA)51K06ueAK wB2m;(mn.e2][֛hȢ bf7+!e*yoY*BEV $BrhREXtnY"Ce,We鐖z)ƶ8~;!m!m= Ruܖ钾twU3Sua56ɹ8l[fh|}fLFtfmک2lij (u;j֡ ऎa uf}x8;Wlp0`&Pa ;:s)fP-q)fװ?G:.=ՍÚ& ܟfu `R|g}6X'y"vF! &Q״f` u-FR&2!db& r3uc{WNCtĴ恇H2уgt1mbu-s0"m^8" 5鈘lșɁ]abϷ b!33dÇ2lo3 IeĠ#&5_#1ۘIfŶ b7A~' fq7`5>e̅cw@%as.n@q[KB-vmzQ"<t2Mf:@DMaQ}0*J \edCjJ6?ZzS/TAqb1CREZs| ]u4+U(4p4J. TkEו#7n٨#zWu1^Y黣 хћ~ډ#0;VO7;Nfd/M$F 9`I1t4m#:ֹ尿68nw-d P]κB0kspQ5 i0z4PL\q 2D)."r;[= ; w7@1w}\9MNSr AKKAWDb~5 HmZv%?2:8vpL9#KؽT*F*vv&zݷGM N"g}jv# @18yਦE:1XZ"Ikqw'w 1\|N1&Q]J*lؕOc7h'08_(5=S8:2i+08^j/fqݠ R^GCO^PN &(`*]8ދ8~h7oW2஁ޮ(e_Lݖ5Itkh= w ,>3DY* .X0a҈"کdXe" &sXi3l}^ ["pπ՞3Hj* 5Fqަ7\`߀~` ;DuF|{zl]s?~wxSJv=(vrin&M<T`uŃ3t)vxAWR;-9 و<[mJL ǟ|WA)ssDs)zw9lÅ㭤!ny9iWmwzM6 e&w&1ϕڄoM7ntbWx[̚8P+Ciq <^2Ѐ^4XhG?zĨ"?E7uGq쥰!8"e!}_$H1}CKWs(!tԙ;,)0ziz ySEi,Y\I1"dwy%%Dʐ,`*f90=c(?m|xdqqu{8bVt@mL( ]HH Ê FDE~Ef%ԬA) &K+R08 |,K2* 5k^ޓكU,XzFo):Ӛ.T ˨^<9*t|,BVBr ֈRX>kP.=45˘ƃl *%%+e#2 2̎L[rtH~X(cF2e{^h͓ReI9"dԵ5f|䔸S޿ Ns̒zg"i1ǹB8UrD#|VE\*KsYu仮7)_熮\VsX<&dFL郱,J4K/8gc }24?=˭DS6 crCI)Y6WP sV9l:XaE 93#رӡ$Yf .K$wyVu;Ij_&ȔœO_=?^) ,03 n7i.mG=z-J?A/,a$l7 v/!TAdN:yI@ǝ/F%":Z~ۧȯ'ѷ ZƉoEdO'cP3'QEGGv;iX- m.U *ȩbbyT[r CaMQQ[ u7j؛ wfSE!8ǸRWpqwY8`zXk4dJ@ %.*j {臏%jIi`Jsu#B9~BMM͟OKj}b927qCmhSnI<sCG V,u)Eb.w|f#r\ghZLR`$GaX4#'X"7$S|QƉ~ڻ)F1â.߅٦}T-}5F(WЗA"bn课M׾(sQGoo*gkB~؎<;T37lzJvyfOUqSaMqȍ\o-/c1+`+q8mVuJ.ѪװNOn2#vX) &G*X E]\ճHs }&m fu/:4`q e D8b Pݿĕ c~DoY c>^Ss<^ E Z.@NcC8b^+$g8'R_i玪HTߧ{ȅm#]!V7?[ _^Y0/@RjF"wNEʬP;'30$XO;+%#o/իGw {.ғ2tTgyZ'9ltfb Go/D#}Kgʲ5Y-!\U3~YP$̤(&`L]vtL8I!32_3+xLܔ({eCP# mI'4TfW9eIblSHv(jD`eb0i"X 'A,Zv^Cf*qvx P4bEk ؇935Iu?M )G!?4l-Ӊ; )+U|*T( D>B" 0 Vx䈫?IW sJiXB/2ìR-ES LH; 3:DPp`?e6A ٷ Q|TEͼ*9`VHz/vDFd =x G;"9ي Wa|;qzI0Rngߕ*G-=voB3(Ѳ-(zi}WtM5T SnWK$vVAV;vpht'x|FAhv{ޟ]Q;|~# }ZǣoXj{ȬDVRB3,iߡ}븨\8 xꡟj"WkyЫ:|+Sq`?Y*Bc ?TN=@`Ob)6x$ H 0l-FDgۖAn9mQf}3CH^`lV![Dt]T󞰼 icHQe&lL%^h)όURFpNZNo2`<3TE%3,ZrNc,2@y iwg,5[l1<x`vfCͧZv hMj" ۃcYcD /k^&r{M^!.l1;ly:C[+y1]cQAjۑA/%S_7``mg?l[;VtX9Vʦ.g"U&EYɔ  ֹVp*D(H64Ek^+΢Q?5zsC b|eK8uir٨,Gpұf[R3n(G0 aKy2Q.a2Nc 2ɭq}t~;mcTӨFfekC1nk 85u_Y{PlT!-э#Ri;硻uc"{4-PYĞ۰&~m6:MLzŤ"KF-AbAt9iՍ}[֒^Z$͑.=βO5pJ@Tdh!2Do#ߣo˂1!4ua۶w=n$l X~_ cH[b#d9bLG3ž;^?kF2O<_~NZ?kX==qN~{xή5Ϳ:_-|8xe~euөa'~LϿMGyoiXw*sAȟ{q3ͱ7ɸޣK~O/pg=|b7:hg-~=]\wp_xoǑ=2&OS'Ϭ3ׂs\4,?0ZA7wo zb'~(=41r,A~K:E;=T}Βڲq-LKX*&)R%w1l3OӔDV;m9 M[bi"[ìP>EMLE 7;aİNq%1{Мލ@3\T֚Q4gAX4КwIαKZZ52 .PшSF2$*# g~-Cp#S>KT1bKM&$?֢T)y;"g(ńUh=H:/n?ZNu %ˡ鐕E0}`:eɼpzL䖳M0bسfj&Xĉ>n5^ cM5pV "38KZgDuUIgWW "*W*o(0T]8XF0-F+I|$)O{89ܰT}dBMu"6F̫[X@-R΍3 FrUjLU0B Icc"\0)jSi=w`KG<[нvHbcxG]pdPM #8u`0ky7pƕuD;'"3]Ǫ6RdtbVݳ-H"ʻV4}Ј$ʱ\M1c"Jz,FTp Fҙ/M#Q rPTkSQ"$(c];='A}Iޯ4XݗB q~4F6Hl7?'^#bVpv=W`I VM++;D׌ ɵjE i hu%}4F y_5>kL+B&EE%'uirKXwIhVi{ݣ‹Wd:'w,1zitr7[ sRkJcֹ^qґ0n "Fvz ]bmI(љenU]h ]zLQN *LvYQ'T/)BLЬyZHЋ0VO$8JYDu9V 1ZUt%%#KT_Pjlj3bpD[tn^?8}/E[T }_!pR6=Rf aH!FHNxYSd b$ "T\Q%Hb^/Duξ{ܵ[G&Mܱ97FYM#@$d2F$f“\@ZEB4sOpɴyzPA0ήrj2Q5n|xwbGdKB7Aev:]?g>uק̬q<蒘U N(hE02RZrs8P=>\! ?6=;Ѵs662U +tp,0V|-ջct}jM-j@ zfx2w1E<2@uff GFI>! o=v=6u>*L9?*rhZ9RȞ2[pqF[f<֎qm]; h C~b%1bHzo) :18.7&E1/}a[gdFұ/KA4F׉͊{)%()Y/W&lX1w2,%gքiٮn#UӃ3h /M{KP[RҲb2h1):]púC*hXX 6\oc?19=Z3n!RZB2>^z/7JbMs#Q- Dxցck0F'/lm aZ}4`Ԣ2ٳH~h<\.Up暙3oݖ?1%Dk 3m՞וՌm $ʃ'f5w2Sꪙmq7^ͻk0䒯VS V\sͺNwЂ"n0`*vq+'s݀a\hU &,Z+R g#f6=>'VjP\5M)l9Cuzɞ>liTcJ8m ^u2es &Fj![ ^ʆ!.JLwIs\Pj}[x*S8 iȎ9Ţtgua .y: 8e+gDN5Y%ӊAhߗ&q07{*Rh2FL@ʹ lbED3?~HwGUԜOcD-hXK_jjo1Ȫ:Bl⬪tvDྎU[HF#kI]Nà6bу&vΜL7뙳 փ;rq#-j͹w9|0x/R90R/!8{+U SLjyݐIG]s# *m4;//$nQl!3 [KLpR@x&Jn^R2v/ۏ$J I\ʗtEt$m\TWټ3%OhߘiE]*{γKZL:B cc/纴AnC\mA躼F/fyBc"%&Ct?MdåK !{mݦNxh{7bw]':w=! Rr=Koo9rtDE7RY. pΠSvQ'7N $Oz#8. ˃ޒQ?b?R^&L4gpϥa bcQ5p(G󻩮i{aޥ1xz{CNYX]f1D br`1=Vb^|ܐ0xfj$3nV_3 =yCXq$YR<͗.c²Ba~uPXp1v)1ӹBk,ia8Tdt wHpڽsĨ;ġ0|XrQmG L5m'`;%Y38^0I,2+jc: æ@5Ǔr4!3>-{R|kl[Kc:+K86kx>WmQw ar}Dپ;78\kϜMg`r-s *lX6̑vifPDaUH;-[o+ R]tDzIDeVBwF­)idH:Q# ^.,0Ɣ؇o Œn9~0\ڊnKE0kf?8$z2 ND Nav gen=A`ڭόӇ;%cp ʤH,~J| !k.]>~!n0'\89X}XK2B9m3$8wߨ]zQ,^~9 qA!r*(R/< N\m:c;@ b {"4T#^Ig#|nUhxK&er^uSW=8VZ(s$&lsxR l'pizdߪ'WN pBjSi)1=i%k ;ICN6Kv2k$>k)jMKժFǻY&藯$8^Gvډ!!5]}]xV`IٕUWCQ4Ʃ0x yp(R .bRѻ (RV84F{׺ȍ_!`sElN#f|Mcl^ƍȒW8)vVKG[նVWźHW\[-I6 tBnURn cBD*[nP9p"eK&`6MAW??QNy :( s ҽZ()gƓUZڢ҇)nGEŢW /oֵ+azf {ą(C ozn“w-%N=3[7-(II] ^. ԫ !F jt襲4e3̎F]kB-%S'pJ |"4| QvbP)tNI=~;013\7r- AhFb;j Sj6EIX|LD9v%-If|n@:SqEӔ u/T'*\!W -Dk b\\? &hdl0ƟȎAwI'ф"jzSFZv5f꺔4ߓ2C$y(HzgNt8Jr=YΠYB:\;{ş:(5 ''c-KQ\[|ep#7oF٥0 oOHT<8>J=ǥsJڋ52)ʯ.?*z2`xmz9΢E]ujOtpxw.GQ^)%OwU}rfUwADڹfWd܈3Z,R4-RdNv̮sE~4mqZ#V7!"M¿P[0,_TZ/GMz=Ӽ'b77!2q.7N'Xsm1]ڥDP*\tD'*bR&8۸M+{ʬ ?̾UNWtߝ5vsY%ۣJ<M#99)tQy $kf\4t0B1=oMˍgΤB۔9Kg´%s/[M2$HZ$ZY;˰Ag67QHRRPj}ƖaZ psP/c2yƊǬgӈV$UWrZ}%g=WZ H+<iGvi4(iA62J+eay F ;zٜnƴH7z 4EݰfFɈx+7мUI@Q-`EI"iH֤z@E2Q["UX%qm~LtsOJifaNB1)ȩOY!Ld`! sfJYu (] k=H\[]f_v{T/wֲ?P2lVOY}ݓ AՋRo-]HVKPHS/aPě7w TOv@ِDcb);iX:}sϸuOf{]TtĽLQqrzŕ^DAI-[6*#V9+Zo(QSv@HzMN6^t>~ M {'':w[3K+.fRÕǠ1ǘt\(7WFƷă'p%l WK%_}pȟX_zbbXBXް"(#rK-ߖ|]}G83{eKGGVoIyy dKpŶ(ۖZ%^ .#'l SN}1*'Btt} auuzFAg7ÉH6kVp̵x%:@,5Pi̠hvh Tmj zJ2ʁMyɦ@3JC_z 9<{+}rÊ8~ MM{_o4cЦX`.sK"W8/V,{'uH Z_rK/6I֏ݐ)m_WP !"pHOX&44٢I)UڦKOOn"O$+]tU[pA> |n<ɗ4P EFe=w;_]2N%ף'G)EG)18<[Y.D::5;Hd1 gl`J ~#_y55:C'_E{a] ߹aO'9Nv7K_]!5QAV`՗ 3j^-c 尟o2gh[ƽ4tpT'PvPR;aZ0zb= JoH7%ņ>*:dUߛ$k}Gb{_"tZ?B9RNӴ@Qs&2LRդEiG+[FTaC3ؗlX|L+58_ݭdͮ+-1H&-gkHF5@ZI2_iYL.?4%Qn+\9/yR?T傧%!&J{{<;rz1b hWR`uBq I*xTR{CNGo"M/ji2mQτ7F({Qnjۘ6JF4ҵYͧ <*˔g̳tѫ`ѧ5y.,'FSpmHiU1Mf@U-w3ǫL2*xr[%V&Txޣ'ԺdT+Ƣ`cHO@@ *R{kHpT ViY`z`@>7ȳP*@=syɭ$^xr&1F+r%Ίhȳ+ړ+TJYN{m|nENKVƐȣ -0(oU@Z L nD VMⱥŬ 03TCPk]F"pkQIc!M1ւ\*x \M\<E6}c"֭RRh^T輿ZZq.iPM:iA,,ݮ0k ƷF@M /&> r5aFr/sâ4@Po C%vY&&DiւiWFM>)2S&"&&V%81`%ʒ$Ś2$_c8>0UɦU5G(B{ &7q,l[JxC=,~k7T%UH"vn(K~nt[)r(ud(:j^_2S-K]`F>'I׭o |2U:t?c3]'J:|ڷ5l; PѠՏN`< ݆Yı^Z'\98 VF"Jcnx1h8#8 iI&:ELFWn7H*e4qcc^̨ $Hn*/BE2%Iv#aƖ|` E,xE&\P&iCPۃv/Zn<+.oXviEGmtV\zӥg:9zhLrGgh0?zJP{eiKTF MpJT p;"=UDca{qNb1v7".F,$KA"a(p#ܐf"qLhXGtKR)RibCp҈&Q QVP*eHƊ PrlI*-[05Fg$9qDlJY"3%<:ãHJːtKba^0T-eS2Fw]2?V-Q%DɓvE# d#é%;Rk>0oOSt>?E;ppX 8B (?N)NL=Ezb 8#1 }jo C*"`pN[W MhA &iЊEQRJ2&脺դ=B+;,p픡X|`2ΚW^mz.oV*@wJ`0TL'-+q.Em3TӐ WˈfExlޫ.܃p^UݲeEҡ--*n*5DxS,x[fY9VSɼs) ?MW2UZ\ݑʲ#ʖE?+ oF?y0镛|N 0{;DOt1JG+\ۣCH,eujL4S |7^.i8:Z!xgqOC!ރú>%5mz-8dGbz 3َ}v m/kyn:lvu-4SSrZ$XHhmڰV<HHjU.†f#3?wwWO`k $ms5zSK ob0ש\}qOy烕~vU>psGf@I 7%3up.NjUV`Mq.t'옖w)ԶTijnbnCstk7dU6}Gu״C*o]@ ?^>R/4NMQ:v/g]YÑ< (ҊHh4IF~(iQm0wjt9\4mXE?GihC/ ʆupvެ΀CEfjw[%t[ZY)Ol:Rj0D#[v!{UaZf娰qJ TQy8J&<͆}PƖRu' r3POj[j/d Y^{} a'GfԒQZh<7GZr9ݘZvE6'.}E\[UHLWWʴW{}yI߽`[e̯Nf/f_}(pQ[>Bg_:G^{l4qv5aƣ<nӘpiyϚ>~(fFQc OhYnr{h<"?۟.~v^I>+L3+h#b8Ak@ _b-{)^pҋd[N .G>MA@NqZ_-mXߜyWMJVls;'Ѡ]w3P645Uy4W;x]'vOgsgYj{vڇUm/Xq@Kd`; `Uoţ/c$7pXU2.6єR24ķoZ ә 0m 6k'GǗ><Z ?nRCD{` ZPZ.CюKǥ,)~\<|A۬=@؆mT\7n W'KVdwc KZꚺu 1ĥd0ϝ0V}$Y4ӄYMas,9T0 `2 !nD3Hfq3Rd6qjS`xJ?06~pq !HAR0| 3BX2:fVO|5B3?, 068^tIɓHže^<_?>S+6!u&:8κs֫r6;AK# ?6"O$eE:TKV=]Yeb"dUlC,8?^fYw{=mSP{/jS`ƚQ˛mVirc9jZ>x6^EDvߩ&&x> Jj6LPęj l 0A1)*j Ԭqak*v[#5F$\D ݸ 7?JݬBhn>7É:"b_{vIMFܥCڙtSV?;BTf*?hxԂ]u PkPcG S&M<yռQ -AV_Mv>@M:vyͤ(R q%CLѮ 1cH!ՄeLH[dyzuY^ X^&tkjݎyc)ł|왺>c_[1Cʴ{o= X֘= Wf0'pW7Na`jQCVu&+F '=ؚS=gz f~B47&IthT85:4{zb'Jb{c&cGӥYcݫYR kB,-6>/Fm,mh&֬ڦ^ ?OzJ/1R~!o3*;S\\ wQ[TaǗ~eiJjf .8oȿ4,iӉ~py' *SA9[5_)G*}ыԺ\pʱv&scv']7{QT!u =+UH+8?|j*?g'Zl9ծ )?*ϳbxiPYmpx5W+=EA9Vb;@*2)_;N|c<kRA1~{Ma9U(o+7ewytT ܆cZ '$`𶖏|𿹱PhVE+M-b(.6~3oiCt|iwOk0z}vļ]N6BOyQ@P9S<"$UrO`޵Ƒ$"LNFfذ%Ot-3i敹n9-3@9A7 <'# HsXM12Jg;_nht52|{c26n&bpJ಩ 97SUNm(x48]r3qnOO _8\|Rӹici? iT֡xlxh{Ůq7!4:`'#T%=ۗK ~qDH]_i3U2溵c0J U6̨^ +"S,)(I.RN'IV]\e%MīF .Zݔϵzów;WrbQ8{od03z#'U捜5>72xڪO)3D'ZV)N!ʜ&J.uDM^A\r∦P]gܙ*; WeI]EDvY.gg,K_H֓'-P/.ytW7z%Hܣ.ٰ1CjHFo%(tAe *jipDS"0QtdC՟?TYXj 5<`3y/L`*[#햁i 5= 4R>[HvWU)T}|Q8**V SeRgRZkG>zXF=[#47eU<{X}K!- S$^LUzR@1z_(TS'@U.@ 4׋i*)yVXLZ^9z^25T@Cz{^/^UzpV^^P<{k)V^n|9=)WzF!.)>*E!䕿CUVUVW=KW]RTU ^1b@*s[suzM!xUAzPe0[stzMtL i9({}>|Q~bm޽s<pSzY0x5;_kpd36 }y.?|co/_?5>}݇mt?6nJ~ox.Ҥ?;icA!F[!A4~@ņ%@5Ah>Ǧs|n3;K\Cr<~ս>4mx*]c*Et}1*ƺů遭45FE0IH;Ƶ9<5/d=DV[QE3}LO &t믄v.y65iCzwΨw:1 !'yC!BCBVC@_TO'l㭱PToWتn%BC G :99 !aH愞 EP4m -g.ma>v9وN1i&攍fYQz6KAul2> 驆(/,X T;`mu_zXVBWN9/d5@f`@kq|I)pΙ ?!$OzW.>׻K%>q @&# aK$9\2/d5\ҭXdKFHS*\|Ѩx;y;*NLwkut27C1ȵC{WƦ}3RҴF=1TaOsY$7'c=GfY{cGm[]|D&`l gfK!xl_^+_^\8)lar~#L t=&OjE|,Y*_*>c?Z6c5OzqXDv,Kd+=v1T0:xlm%I(P-t; !иz2tyyw,1ȓ.d dk YXԵ+#c[h l Rx4Y)2E/fmZ|3f,2JFSLӴ13iN;CzX6@DQj l*jzk&7{ !ofsO͟K<i~1zus^QTG:\3B\WI{I>D/&~fCy:AڑjfMfH2ܡ! KZ,{'Ho)i,s#a;" Y ̣V@t ԪVd_D]4h$,}pBC5Vx&VVcebt1b ~#@Y¼p+pp*PͲby-V0s Bh`>+d50@v/mh+4V|/+QK%a8c @yD0'd=D vRcUGE(J4%v2+G5I2Є6BRO XiDG}NX!d54r_ jg}t^jVsq& ̼_k?V)㰔QZ,esp1琜YҐ&y4G RI㥤q1 M|I!&"Y!!-١.45sNi^ Dtr c2;-&@Hۤ->SeY(# uܡ,#QEC)‘"ˬՐDu5$O$q((δ$0O}]yΠ !! tL6Qe\):tq1|?_}ɕ9!p>,I]N8Czg0(5\%@oX1>AQV_Հsnе5q0dW2DhO1."Y!! V!(šD(>:B 9 !! ~ˣiQ-aCFؐ9aZȏK;02$Հ\ŽY%so./k#:,x!%d=brMrN/,dam#B?p޻[IFE 35BJcO|nR%j,Rь ֖.V`HP"+`E@d[EnK%&4 auy{#gS/Xa6UdǛ߀DQIbg`jFP`߀kCX (RϿ? j{Lp,G9h\ԗ"ƴ\"hHWQz^Qԗ(fEZG7r6* AU 8`u7`\ظ &G&ek#,ˍ*&-<,5C҄#(|]%<,JhlT0~CMM~ҏtcZkԊ6b%ü_Qi"Z횖oPd$0x!wiḦm $H}qŐTJ%0h?W!AKp+[9\6Rq-<+'O-Z!0Lu&RXJZg:_Qh])KUh0pЏ(tc)4ye:F#6=+4bF =W߀!I*[9PT=ȧ<A U %[-|;.U Up,PlTqTi.d] azU٨=HwNAwf X4cY>ݔ!" 5HOud=UJ @66"0j+TD<ȷ֎ -܂dԲ{&m&*ހL-+I: ^v:d~kB0Q.Mo q8G &~6'[RnQm:%z=XHfS/ax_ͅqg [K9pTjٺZդc ;*+>օrgw5L-c9α [cu-~B=)[ a̶3=LM;%ˍIȴI%__ޯ)z Y'A8;zpޟΠksZa{ W@35]u7@, ~D n'2x?Cbh+uܱϛP;WA?U7_?: pKMUru|ג/&[W}4W>݆w?ypo1\z4IA{U֋Dv6t Yj?LU(&`ajwW[qڷ][]_}; S?Vwfav@N 7Qo[u[uwv1Ⴁ7ۊX)%(Qc34_+1r6a2ܰgP5}F̥CC[l9\EªQc`S_ ]b3`4(4a4oӃ38J!Q^+Mlʖv o e8dmփaZCze< S"aΧls.%S[[9UAʄR}-\=>Ю}&v0ؼxw `dfu$ϒ-{/N}J/YEV׷؍<|W}?܁d,5ʿ5Qxlz{ܯR&3][? xǡ i:;?ךqӇ]@z360o_P#!& OX Sx >vϲeD:`Aw7{[.܌oFFɑz\3q`ΏV_]?tp# /n<H/ݵS~$p Ba{!e0 f#`YB7LF@`0%<&2HL`)lgѻ0unAf+%u=HuZϿJixLç[p|~lbi6u}1̔>5k'v$5}{1#~ߢ{o'.$s}pcK"6MpZgѮF+5.:=ݞnOrԵ¥kĝe$+.3f#i@~ n~_+e`ɕ䌂 1bƌ _[jn bۦ֢q2sM0}(=&2I0[-l"URH.oKt)G9rBBùODW0 [:.x}reѷ籎Y<'C'UEܣF Rlo\2 OZ\k̸>պZt:L8u9|0 #T J=dFQJ!}!RLp(n€৭ˁtObN5r d;‚zLN9S`$F(v7[zECmP! ua jFjc!9$R2&z>br" |Qg>Z9.l/uCTYƊ:5j=N)Em]anev:ŽV BŞ05Bl^.^C6ĮmC`XmK[T/#Jl֏m֏}~,ZATڄ%]+B XQּQT"97/B/$~dQ,`u0MQ wo,J*jW iۄPن>4 C׀5Zc%Vs,x`ՒC9wk߂R* `k|ȏ[2[2uhoznIs`c8vɓBLcSm|v29l0 OMEIg 竼Q1粅3ϳ(}+^ށf|9 K(^,3%mqߖyy%;{3zՑUs#X,gڱCLW |lN2R8IWS"aSmGqeM3Lj`L%Xc-=f^Zp+$jrLX59&cT[瘒D&d*Pٔ4"X2rbSҌ\.OkJP^n .!e׻7:WA$9C$2/"#aoP0C0 D#$TP-J%q)I?$K;u:m\O ?aC ޶. !TZ4x!&3݄dZ|6}~=(% mZ;CE)iS?F7#E!!%R\hN_[B8ߋӕ}p%47v scs #&_uޘ{xW >mD5WD^n2aN6#X{O0mB-F7BԧmjN 9].tAQRB?WP~^!K>KZMolʝ\HQULd]JEXMP@&;| g .~Հc I9Lf c0qrI<ӌ`EirqR'NzJ'KTPKO[':9fQ 6JԌ$6v6O#(]#2'?J׌`O"ܐ\uڣtf+Q"܄cXЕ8|w&\-mxY>#^nt鏡0^ÜwaYNQssҸl@ھ$P-e">AWy' D(ބL̥>97"XSx3rx1(/YtU AN"L%̴r2(3?7Ȯӽ1yIT~XCf&]m^z||i g6[{lsdvu/@ng9ɧcrgV@x38?SceSΈQ%'q>ADC|٠\N@ދwD.. ΆR* X .DRrQOhIP,tcm,5#vMFdbMN~!ħ i&Žf )킉.nd|} Â"trd US>|h6Z!cּТVOѧΙru038vƽ|5yڈzGj& Ws}#=+J4UJnؕ\xӇMʺ讪+np!-hJn/gA_ʽk06X%y8([ܬ'ψGEu+T~1{6-9/0-H"6?QO҇Ln*0C%i79]koIv+| Rݺ^`7@H6, ˢ#ҳ3)-RY|U=NwV*P%H-9Ruq$ EeX X}o3tIWV%$GBYŢ۴Ŕ΅uc|4sG~jhAR_Ǖ9ils)jSzXN 95QǦDFj8(=V;'ZWkS`8$RI0Jp-1d]v-s,r^yjZ3&ٺfrRt3p2I0$VZX(-FR k1MkKfhf NYVdɪrQYW=M(f8~k8d*!i(,h4ǰ(ͨ&]PJGM)5LFar@p+&p,yrh..5Y~{E*'@x`htPIo}{sw>f0eKd<%C$d[:'eqZIWU&jjbu9!h%9Rlf-HJQ+ޙ'@xhԱ4n "NXY#8:Fӏk9bBZVq m4Μ`MDUaqЗ Yl-uwmSq%"`3^TqSVM٪\ 2)\rUl,` `֑%EMhtGj`+ݥ$j S2](!OFl0`% a[PQBQ8g֩`R\7Zb7kx(Qj1PyUc-}(c$ca ybM KDicm54 q'Yge0P0f1ʼnnJ^e>0\cds84yk֢RRmF@dq{X DW #@HlQ< R atuc=f +(WJ%cjHd < l3lBE+l`'œ`AV# JRhUQ JSrʴ.|cf2 C@zWs J+!z3|9# e CX&@C,CHhPT&TIDmҬZeJ|kJځ A'b͂G Ä| *drI0Όx)ED5 ,3qhd_ f -gT g2JAQŁ.є&ՠ:;J` l+:C4 9'0qܳ. RSBm R)NWJhZIhz9FtY6D۝BAzp7_&M2:(|_ݢńTE~ƬI@sbh#(*KN.` /0-j0`wV?xTAc[2V=V3w m|ka%a15BW 7>o;@~V2z$]4Kw20:%N k!h)dA VLta8ѼɷK:aSе2kwPb fŬ>&ް[`%@_.$0%n@R@")0eP_Кf&Ab9*յcAYx kPRg]IPS`6̬D4K& EK%1n9œGBjAaK7#3vElYeFhQBÁwâD6H6%T2tyXدչvY MeBoit?|7SAH U|6pڠ@4XA¬ e- X wE=pe]?MД nFI 45.Qp}J*C\0Ab~?zp4TN)3*5ٝ.CC-2cR4PF!88M]XƳuOC?ǫqo5Z_>vQ#R=׼g^oW͋ \?zM.wW{RMǟ}>^2D<"<vv-&䡸g(ݜp䡾<kgyyyyyyyyyyyyyyyyyyyyyyyyo;}1-)b\:5C-C}y( #5P#5P#5P#5P#5P#5P#5P#5P#5P#5P#5P#5P#5P#5P#5P#5P#5P#5P#5P#5P#5P#5P#5P#m,,)^-'e" y0^dv}eWl7=v~svuO5݂_.m `0wN]~5- I녀^C.l:)X{ X{<2  Ac> Br8V\XlBZ;˝e)M!;,H`s`Uw`a7"[b~U7gooeg:ϽAUl@_jU^/.7p-/+Y补y!u@ Aؙ[5Y` 3!,!N`9)'.j^Ȳ"{v8XQZѳ>ksGȮԛ/s}:|TEZ2C T T?-w_ͻ7oͯsn?GGaˉ`bBz%^ۅ H8ms= ;(Y֨Y-Ua9#J)~w㈅ ?bչ".,w I,/ed]L Y:9!Xް=)f!`9sڦpXOė9e:,:#pv8X1dh!`Ety5j2V. :lbi;>>WZi/ߞ ,WBHktϳ>W?| *=#k Ǡ!hu'% it$\/'}R1hY)u'%Y=)OJv<)1SDWUR誣}XÞ20+oо$mX ]u%z.t{B=w(o 3UUOsAWK+>p\vO>xkoLW 'm8]tŃ;ԟJDWLp^ ]uA=wJ}O`vJ ̠Y]uNCW]ytQPW"]Ԃ b5~)tgo衮E2tovm+n^wo=DClEw}v]#>|n C_ol)ɨw@\6wnI>v +i}!ymY_yxj_e+oHbeym_~ke|yl/F:^ QYmo^#unWۙ_|7?}7oU7mrum^}߇jWjzgx]]brwd'72-óuUO:r\^=ëOG|(<k۶Kn $;JO<<1{iju$7SK*URVE*+)xS26*X,|0 M׀?r L?ΡD#EဌyPnŘE˱!KWm̘}:v&wgj7[~2 #ƘC QlMS՘e*ɹˮ.}+[㧏by_oG[b>q7zL1ny|3jBI})b1㯆cإy07$cm>9QlmˮM_LvXMN(b=9nEuǑ `YVS-h)x7m:6r -O#6.!qǰd{7ȭuE\ KP$KIm1)6xsw61|bXE&d#HXPv?}r]W8]"ͣ %yD^q-Q#!8:,d`˃3|\+Y#V!Mu@ KVU~qd9G0,W˿uYNÁD y1\ȼ+*?8rHKJ'ck|hp^ۦͥHPA!nSDM1"RzkzdacdLޮ6 ՊR*P6[<9&y)RsD͐6DzYؼp7GoxK |r^`fZMGRgCV(sFN`P׹e5-YNdkTbfHFs eqB8F%8Z> rtśk:>0[Oj%UIcB#!;ld3!yțVD-F9<"ыHHvA'GBu UIk +w]Wh$Ū>@BK2DMֵlɻ5F"ԯ ;le5uR` {<[Y[yd2eJ\a5 |$kS$v6:]< 1Q1e7OU Lg2&?sq#Ĺ@dvxllHA |@JUs"nPX!4]ZG&=S'т'.Q@hnŽӅJ0CY^"Ŭ5EM;݅e3S*妨//3}j uiM⺂Wi&cm.7mS u}LEy:D WV\0.q|^9Zm^uǻKmHKd1s,Zl!8-JG @u$+U|U>̀8{hTU)OFUs͗pN_%Uq\'ylVyӼhu/W\dLemoY[6k fmӸGV ]WnиzFNX=N(B&ǒ7좖1؀*2"eZ+:;"LOQYo` ZZ4bV%|v_@£(E9p'Ia/lZqpt{A#7gzZ͔UjJ(TtL!!,C aYýD( FwNA#,&7ك]2Zmr^Lg"r)6F햔hmG6[${DP*56eLmVR81U6ZF[ѭ_d).U3 R7("XjqϮGEρT!_t0)L)(g4^6=4L?G7rSvo>ۇC4 7$~LBPJ NCS1b8=79 ˀG3>?u:H+Nd3qäΊ#V_Qa^YLXهh/cBFhCւz"l*q!qJNtmTg7!#O1)po@Ų./ij xBpcE<@8%&kpǝ F e|25l 8xcZ,\}Ygx0ML\RH`F O9cQFé)~Yr1ip Cv&dZzn ˤ_sV.q{Czm ~q||[ԽҼ |&Z+Lp05Yd oU !ZAyui:"^/,<~ SHn`> 蝎 P=閇No/|u8!Zqy}w'œ;8c;8G1n.[N.x}cGi蝤 4c$z{;{DG@BO>5|یW A+`kv6wߚ|qBrC7~wظ3LʹHٖGIATb98eMC.W?d ';x@YMQ O 1?pA]vXv 77 AFO%r:s@9΄`lePgJqn%An:@}rBqWN^| wt/4N%O|/'pG.u&w=zָ}{?'ܷ&T~ʓ:*_U,79<@t vӸ RM[o0/[m*'NjcSf RiաW^?:S짔)3#ⶐ)V*>}sN aA0 DI;Q6GMT~-}#pm6N_L6,Wk^P*alv7O_ Oc/ bheQJ"E%& n+Oj'үIwXK~[XCmzs&VӸGbEbrK %Wt` xz\\ၭTyExrT`S(9rXnU[:wҊ3lQEIqP%0AڈIaT| jҎh%3GmbrlCW~264pWfS\jZl9/5F]:}J9ZGnŘ­ k`AkH]W!~((ab~ĒbLsi㊯UӰi%nuj$dq] lv|\- AM]8}RUbMiggm@Ө1}:QnYbה|7sj z{Jlƅ:h:JeqY` J#V/]Ƞn{X*Z SA/ a* |S zފKɄ2H 8SMV!Ȍ-Q&!Cu88|,'l]?>We{hFUf0\̚QTof9+Q#byyK٩NU+]pBI5xx~^ !ȳШ rjzoԪ{gv>|^:A2 1 %NkdD.fGNH"\$Pq>Oo6ԬلߠA9뚸L6[fR_vݜ;^Д/D$ yBK"0ث@xKწ8 lٓ[-6MŁDzU'(¾k .Nvt}2תyϚr;go{BGrWvn۞Y :T@]"q>9g*@zh\fSZr>ܵ~/)ZOt_ k L0v%]/LX9VTV0,X3mV:W(Ebހ [GA#yԣ4 5JRE1c) )t漐0CK() E+ f‡w~> x tD{ yDCV˪x `}%((w!;cD,y=$M`6)8µw!:e0;\XCvm1+ )~Nx+.kJ'?lE责KJtzo4GJuF9Pv'BUOLhYa9 |$)n|{ bt:Atf5YMX6 =3j^^@ wʃM/^f~|O;Wfgp6{\ų%8 eFUptjX m&+=IHYt=kXNU!vߙuu6弜 qkЅ_|Bx0#/zvn!U 㟿_.*V(X7;i)WgG Dުme06נw<{hYa9 |, +Xvk*2vj.Sk2/Uv6SNXZ}9z-I(Mi/#<bR=vqvuXS dWX)Cl7=p-8pt@cY.MMGfGq]]ZUu! 4G6g-\;1\8V^ 2ӊ @ vUx ja6U!IjqmOl ,m!t.Cupu6^Qxyam/W l -ˀk*1шYlΜӬG=4 ,' 0ƑDzo-wmn)AW|EM-ÀtPdk&%}o.2p}U5+ʙoj C[MXzrPjW!<񺂿prɏU heШyusk)*rS`&&lY} 6>$'HuW2!˕4 \ACӭRcpL)ύ 4D G7r9e[61:zJ!*@yWNKػq$W,0LD/bbvp~EQ'e߯(ɲlK1eYw$6UEUXzzΆS]~AOSu ]b@,K۝B"' "m@!ÄpċaǏb")UhOs?d?v0d]n])LrdR/ED6* H3G,Fq* 9gu٩yO]:4qסC )c+J8 amZEn%ZujSW/YlqfjG6 )α/}M#صZxT'S"=/rdEZxL8^ėr:K{ŔB&5^ Qa[4k#x,'AvS~B?\H FPo@ε\2Oj]M/Gb"FHyxp5GK{(VB#9#!/ƄШN.4j"wkWC[Lqy yވB.ׅ.-<& pVF1Cu複r u:giHwa lVTT'lW0ńC NΚo5Y´ lA72@{Z-s=rsJ>&>n} dD#"O'U쏟ɗ 3DŽ.w][M/zрLx{m};tclH$^Hq^=`Ä@mBm7B\lf3j`˖<*āF*qjS|DXŗ8:䤄LPa4cG)ʱq ZqjS#/ 񷧗^ 1)CCc! !V.p`lب`)ر9ŞdDB{1kqt AH܁d -6$cICɣqQqjWO뻵߶x;=mQc@wPWNybaF[w>>/.ZxLYtZϩ(l7LxFpn@Gpyn:5+Z9^z'CǤDŽ$lpD8#Y}`p8+g8^iK'ҖD-a+Nڀ`LF 0q?7)"H+pPgoz~g? }\*r5ʤlU1`SbP715-ȳ9p Sʓh)[=K׆kQC/iFicy4K"._L\>u~Dx<%I?'iuQ>EyN 0ԁѰ<*KЁFtjSr@K/ j MOr vٓ񌔏kwri115%fN| 7 |%E` ?4y"s_~Z{'$c=흜@O!MX@>)6KLeD+P0L_R[5_paPlTI̓KڵˑɄ0)WƝ q02'0FCıQ0%B'w>KtiS);+jOLlGїN[xLh``E׮OqՒ#`^o;ڃ |2F )3)=D$%J!Y? A_S}̪j3q FäE<,SDi^ڥ89-/3H iC6+a.#; ZBjCVqSYt}k9w0FڂKZx|Pr䂬:<9|Zctbv%=AC`mI{ňxR8҄RZxk &C"IP.^UT4W X-3޵s짥sk)y'rbZad0lT{ϡD1a$ԔNHDG[-6,I=Cϳ^OΝ{/]s@:xܔw-ʡzdsĕuHkܷ0cn%_LRZevɨK2[o`0ѡBZ$. RNmv$Fd>\ $mN+ JܸP)ǭng'1#L)jzofGP 45|;^;9|[x|g೵ɯxҊ^] u†a!)E:U&$T1£2\/76 3N9T">9B#y/Q2N\kgƣW ߫;k ~t:)N g3ҎL!7u3xF!ݐHLSacK-5lAjϸT o/N:X8^]8}N&tBy`%P!g> <ӑSX8J\ F뜊X8 a׎ZUqᒺ?&>8A=P'P oCh!@.q ;T{ 9tv .qPtpo1pdkp HH` Qm,cÐ)C , Z$1-wV5jlu\a cvuk1p}vBgW0J5pEɹB.klu^Yu@kC6X(~[(\uW"` *v%+?0mww0ќN3ax bwvcwo/qd$a1"ei*0 34hp`:-2-<Wأ/+?->8 ^K*1%`=kzyN*[x ,>J1tDgP!c{)s \X8>^?i vtB}`h qn&XFXWqCJFx$m~^ـ,)%BnRPL.qjC0J?%dBR(BN c`\swqㄥ7%rXS=A 8F! P.y +|܌])7)._Ud2-_iXFpa1&DhBZ)^3:b3J6WYVNnJeqslQ ~.() XgWKAw՘.3oew2GOV׺r T5 }j3n%_7n|)Vihc_䮾,r UTt={ÞIY?/b6l7_{y,K&o⮌T91S$JEF}~વA- m魇@_?*``0 'J/Jq/HS<,"2DAXOޙ→S}aUf;oyȨ9 ¨h}?ϓ̟dX4G3o"zYDb-ˣIf/xGv;2lv49_7sVn<%˗~{%6lZY1ͮl]LV&XQ_̙e[bײqaX"rcV/sʇ0[(^9K77֛t=ߚ__`ПJ(]Lc@HHr!ow:eՌr? 4efJS7Qz-5%54Z,\jY2m9Ҳ6#q?{B5dj}wKA;Gk1.7;-YWæv'FgKMl'|+] V?m[{ rd_5c-?j_z@XVFr+Uu |Brj|U} z+^Z; Sh$djX#l43ƭuTs=FU]8Vtjܦ~bo΁Q-GVQ#L!Z1_Uץ\(Vsjܦ[9 $sN5[UM6\hWrjܦ:YV[=ÌKK=WjեVi\F4aԛ^Ng:w F*Σ+bi{WumqU2 n8TjUqRs.<*95nqZa nR(HH$u D))N"qu͘hkbsvoݹ4nes7^h%+k{aPS>r~n=0ɽ'/b΃P1i ꩴ_h`v̓uhY,Pm>QՅoVƭ֮`ZtQ!b>#oҊ:@A:eogϫ}ߖrWʩq izkkHHDl($ٺ)Z~o[̹j(_;dwrP/.' *>^Y8 4a ;D2d)ށxQ$b-^{aCP_;[r[i˩qܪX⑦<˅z0Ӝ^2`cݱ1wl[ hWCƯzJckKxѴ咙Ӯiˡ$=_w 1X+qc"Ph֪}RY.t+e95nS֞u2u|]mdEѧ\s7@&8T#Fw`Spo<_!X~^HHJ80 =-H-˗|ݮ;ҁ޲tizh @bĀL'W%o2 -FGA)QԷZc[d"q}[em;cۆFrq$i,Ԝgc;sl8푿vjdn2Y")9zK9)=V\Xg^bgNe[/˖|k/MQ|fgZYnʶp[>^M:50c9R7<p>D—q"j>:X' tb8)*ד^><-7 1dN}tf278݂D5cș18="ƘJ'2 42 뚅.-eW:KP︖/zŒc4Ih+P JUK#Iά$h 3R}CU>o^qB<>?ciO,3sO2iS"ZjL沤ݡ$#,`'d:ˆB8-:XžQ<ɳ3Np9%e^U $N3~1`^uBRm(CCl/e&+sHFd6fV'P+G Qr]&ZĸZYxFIyPу7яf% _,j.G5#HBU޾0~1:za2FD4U4bPva"BRNP trqGW:;-(3g( L+iD)MB}&!bbo4"J۰38}XI^ZT*_01_tƘ8aOr0sqAF%a .rJ#:PsH5>(v%;4=$r:"6Njr c\-/, Hų4l?TN/5>į;LƟDoT.\;z*zj&Rd>TT\\?ޠ-1L(R3sfii#`2b@m9\AliJ\XSQGqصН+`صd#v2:rlOM]~h0dQ+(e=,` ڂ[SVx%Kc+J]cF೼NEbh:9w!ۂ;$DLnHg=I<(^pԪTYXͷit7 ^q[C[GPq;7Lj! ta28doӏfaʇ؀ϣQꢋRYD+/M'3 JpVrta wR*br}l9> Z 0ceyävf¤*9((%lwvyk8} # 07rCLLUfώ5(fsuUG`X R]k5YKVͽYGˁ R2Eq|k- @ɘ1"«'G׮h =jQ8Y/.ⵀպ;24bA򥝹0Kݝ)Ei85ÔdLcLiт$m3)Sk3| ߬ǪOG G.򤙼-jeaw7^3BQljY`[MĻRihp`lHJ;z.ɼL8av^`yfu콛d1NɁd-1lA!x-Fkō/2 38Η]k[ âyd7H.>|qyFi,b.#:Qε2@GߐX&fr v-4G[``鈹 L$QSQ2 n,߽+>br%\ 0b!ch>Szǃ5ә,(LLG\4eYVrl\km!@”0% 0:@!cTNJ.•;q(C(ɿke1&ʋ3FFM6/l􊰕X1GQ-ҵ-1c%^^|f'ͨVՃq]Q:X+@E QY)-T7&k(ZRvۭ{z}C i2SpfPx1Y;MՏ[]`DX^IRq9mZ\4uu/ydL|X& w|?A0()UolZ$N ťcN.[g[ i÷#TOuT Fp0<<'E쌖 Ltzfyo(^9,U+tr&΃6 pqWL|)xYN4 ="=kHFSy#)%ДxYA<,i&<5/ } wdf{學̋`flpTcN2 @U0*FJo{А@!07PQtļI9Li*0a4sIKFr(̔#SVxKb֡R5HkUv~|>ep7i4P¯8w?y>vQ;W-v:>f}_]?7O|jx?,j0?t~N+Y VbrgkXglR|S7Av~jknnsLKo|["#J\ W0Nt? GY uP5oPc5B[G֮2,ϪVX|n똶Ieͷ!T*l)nO=G[߬uzWEIBPmi{~YMO_TM~KOZϞMr݆kp[Oӆ޶ԾYå{+fB>?^t@pU]v*X@Qѝi eU=R4=f>ݠ[݊0vcQ!a@7u./YQǮ%5=UY" 2>zK Z2+2!i;r,,f-|^8#o|/A^O9/q.SЅ[@;siwZ+=ϜL)%"tbN2SH" c,5}'\[|a z_em˚iEkłfK2%l6-%ת`ƥ9_i;rR*;"p}""]cfhωM&ee&Ttm3sԗQn4a.tz7rKp] !l[K[2KV Q=9\%-tt&c#6`6Ϝ!,tXe`V08kv8_p 0-#pS3NՈrg+~{\w:z+a2(e4wK篣[xG09!XBFg. pQdbPQ3c*~>>TkKKET^ ֩'VYą 2ci;"Tӕ!TLi V(WrS@NBӦ=૖1D}۸1g[lthy6LP<ņjDS%bk)&c JAPK6>D@T01J 9Tm%C+(f⊞R-dv5:JxΎYi.əDMKp'84O^c5ϖ4ORV.nl]!{%e1oiY.ᐂݞhRw&k#m94b`'˴{IC b4̈́qi.KBÌmnF/5_ʕΕ㊝ܗ\mUIJ]44] %>X^#3 tOyx~I=_ׁ"g֬O@Vv0PR?2iY$cqky$gcSs⏒!`;X!0خ{A^;@ D1 :Qp>94E ,1r泯FF]|Cx[ ❁mwql##zȂsQk"O#G ȁ_.]}MW .mrWٮ*h xq x룛%=,ʏnKvx&:5Ҏ42^^C%z|H59!fLXό'!\>k6x O'O9Q%f!ȇPц!Ƈq7=v?i@E{?2R ./8H_yaÓ,pp0$k#Ǐ ݳ'zb)Wܥ, ?K}WD/ᖦ=e=f)z_d&(rR( hxbbq(&!V% A;d|3"`0><5!U=osӲo(o2SԺ`8 |ܭo"C)ks~ anK465b2ܓ$XB"yxVBN4 FRr 0FYnej",Em X# )a^xNVperc͋>.Z~ZP3@ y/hQu0Z`˝R:gLL^(s.Ln YްQLwYS7ncz:0gnTVs7Z)(Fh\ h Q ˫ݖ‚$%ա8r$jgx"qvh De9bQQ!.(!C=*makH1"8d|Nj*޸A(WV2E׋UzQv[g0{g=Q߹,Ey+KJ ä"#c )%z8x[8̮Y:;0y0vc0 k@ImlvqWSc\e[,6Kuy$.!ᬺL?K%Mac'E({V ,uR#J͇4‚o;z0pc$QEqYp |>ZMG6sFN*ZfYO:V.;Qqxqs0[QXa6f*̑EEZ,s`ICo GLPeע7]a,i52`k*:I 35¤K8v37ٗ0&/!j.W,7|2[NK>~Q!.BYn"km1WW'ǐ }4WuY(gϥ- gӾbϧFX18^܇2H畛{uQhJ ywAIon0'1Ei;diyRp('+ rp/yN1G4(Xy`F>s1O 8ޗ94ӀDf+՜1Ss[~@ÁkЊ`eö9ϙ{崔c8w|;MEriܖBGZn  ߕi):!-8ԜNh`0TȦ6/\Q17~qSS]vs4㐹ü31 lyM˜-}Rjܷr@e =ꇙlߢ2SauǽjբtqN4D|źTQ2NLK y9\~̭`^LS8Hs(H^%8:Pn+W[Rl|Tq90EeZ1wAء18?8o:O! Z+S܁LqէR[p=F]/h{Q_a\mZV ;+LoA*3М]9b)k(bVRo"nX_h NǞuEc] OU+-7A# iSʰaZL9r8vzlѸ>ŇvVW++2 tt1!EQQ]F;vTKqPY`q~2Kkj$'3jkGҕt ;_Tr%16_5ܷwwhڀ&i=> U$`U`1Mh[خ%).uV?ޠ4Ia&T'Xn(OϼP3z(yw R |O,p,;"rxdf` LJ9V2`=DJ%6NԷ~~@UִCFcpP> = ]=23au2a&G$l9*.NXhS_ t˳ӜԀRްQ2-N$j)ӹQ!'HIFFcp8^ߕݸ}W> ލҫ}L M(1 J?CzHwh G'AbC3! C&s) /aRr{?<2Cǯ('I6u֫<`s*IV!18 :tYfƏ_-^ P&1, :u|  l-6H@.h zv(ˠ!Q0/yPy_ CFcpq2l21 GEGG NQEQB3<~_9֜M9}m7.JI*43f6/>Ua|4zc@t+{6U(E\Ӓ>%.!m1861mƀy"1cG5pkQ[2SLr ^yHd{dN]YA#G7B/}{efV-_4[/qw>H QѪz%3zD0x#cpba-qm֔b@Wtos~'0dzG3kE1]"rR` .0b# &z_&?T,e4Gg%ӎ]\NhCKUڅd{%T&?~?S\j%j9 e08S FNXIzkZ?jHYApl0^%Ƽ)o~==[ȷ*b.14 XgWW2ڂð> 'HщhI2Q&hlXHvzww=e#@\ 6KUIcBF\Zѯa~ȸ/7fwL~^z?KW^q㵂'C"8=Ɓ%y#HETIFu#PJֿm*2r! ixt0jЃ1D1P2n,k퀶e%GCE0uSF+9 Nx)gl#(:0z8xS8 bDP;'ligj:gJdܾvk{)Z|Uh,gW7ZvL+^[*J7@@TICS%xb҇UL8@1\SCL RXjSgiM<O_*TFuSOR]motHomݑν݊6_LJ;OEO߾N27{3W>Kw'E/OץV~gS\\T.iX?6~tơ.tP]ۡ~tG 3 lҾ[E*L=WΓͮԽ'sd}'?'_/SʍqfOp ksixfsyZ()r-,A9 JTP9u*WEECu9B6# PRj,АȜ :/*dz<='=- &<8aً?|=~'~ӧ~R׳³߅r^dig?β_2jmu1pװƝ~vܫGB=T^m%{TO?VD=$.e?w+SO]8_״G; DM2@j>ۚ\wG kœV _\i_oJwD].z/ gsTDI[/iD>00V nΉsR/G_ zx؅X٬wruq1h/NfڅX-_#Yg>,V"ħ3doT^cxPro;5{-vuNI7}3ƈmd¥s&:JHPC Y&QT :HkA|~AȻ߯f/_ˏF/v;x)+" 5gwmk#6lL#oveGR MĔOHzu 2&*J_c0h!c";@L4' 剆CU uɮ䐲>%%%DoO/߮Ay|=W;h7 OtÌ#/Wng"9BZ|?=dil_in~hNʖ1TWGŗET>]O[DkF]{+9w2St~[xʢՕ2NHzV:Ѣ0")qIrfKwigKwjյvѻ/j͡/i$3C[@}v߶f~9Cns|m;\q%xtCuk麥6ߘMKrQ5G}1a1M /[3PBl{|wܾ&LŪg\s[_ժػ_ݝzܚǰ [օ /8<+FԌg#%X Z.s"M%Ix]v\zEny5ީR(|z~V6:\<|'j!2g_a_ |}!x4 )ۭਇ^,3U@ܠk{L?WM'90 y{78O/$4/KB9S-Tl[ZZ5ȞCGbx739ڂkD4`@f`ԪFܢEuObe5L:vccב#FDRIa<֎ru尨y  = '_Dw?uLzBm~PKeIOpgG f{4G}5VPˍlpW] 3~ ;7 L TW$3R PXqW+VPK9W_;`M˳qdz~# D&j!! SpU,1_B |.0,aIxWHPD.HS~ -ɚ<""A(dUU J -[ZÑ_#xLW+ 5% t%5Uk$4t8_sNX 3lk(xEQVEH=FB L*J|" 1+f‰(S5Z猷6{n*yIR'%AYpWdN+5Zk$AoߤգzdK_! xC[348+HHP9+@Q%(8q k$/;+~ *E@2[oh쵱Ko:Jz4$p5i F K=fx=}t4y?|,ONmOMSNfk'ap\Dȱp8E0]#xÅj ddC3M\B[+=|&u @ k4|Ζ8g8^G PÇk$i@ ES_K3o`Æ*K= M-Տ;vtov4)}9D-7s_ϜCzZ*KjY\kp ڐUCv9(u7 -s&[?r:8PM2f̼>P - nZk]JhHLFjK) ct^{ʸ-o*NxNz_k$/n]WE2 ܃YZGWHh^/03%z" EL<MOժFB =`<`x%B,K+=3k$C^4Lx΁skS$7FB 7/ӇS"% ȥy0 l&h] U$Y XO 1+L$gmAug !Sg F*I%)Yj$ϔm#9kJ}r62BvS#xY"2=u5IKDub) -hɚ9خ5!XրVLphϘ ϹD%Cpqɂ6r􀯐nRXkˑQlHm QeƊ'S}I^;))8"$0e(}k j$^s>| OX3+ܾQ翕j9l%* `z=rW'x]틻BmͳwWޓFwe!cĮ@a]Bm?L}JSnhwʌTL&磿|q.OQPA JB#?mJOWX0ΎrQRd,{?D3 JУo6;_|w?h$d -[[t! fhc&!l>&0hg0vjR&yM 2#72xLIK‰"/EYM֋lNM*M8`2=~(hq(˼\kSaR⼵j9U6`Um :HZc+\ 7VD5gL:b!+JکMb-cpeSlZk[8O~76OEG~OV{{۷o=wW.NS: s|) }uA73zj~~ීԅ7g<[<@v{N 'y=4v8[>V@7cZH+~כt}[iCYdt1G?-l)t<~"ToiUޮ|h<f?kX(7M&]x\嘟7LR~4<z/s!y/e,Rue<+"d#9sISD2S6t|:P|9x$.kS0 }Og0`6./ζ曓<cȣI>ʼnp Ϫ]64r]@{~7?O>7ʆ.XKN6?1W 4訷 Fu`cH?FOs:^Ur8L~unekmQeF)+ǣoAQt2ώ7eHS Fg(h~6۔>OOz 77E )Y7?=DV@!w۲E/>".0sٱd?>Qx~mF>֟m8=%0fWIŎ Y)I M,Ajr /1Cs5n/^2 չƫP 8-b#t^X{QL/UsޛWt> q Jr;7Y-^;Eo7{JKwZOodo/aMwv߼!68VPos^t!yͣ>];'75'˨o5lPEȽUn(*o;}ϊʛv`ZOĴ {ț"9ߟ́P]ae!h+({QK6xt'v69˫{u,ZX$.V~տ4HU*Hf#AG W?liRC4+@)cYd0KaJti%F4mRÛջ{&WS]o<?ܢ&~N|\emf~L(3UǛ6~;?P`eO_P]uehqő4; ۆ[&Diz:4}xW>tm \R炧s4g.yJX4 (/$@?4H7AH H5G|N(".h)U46ц#nh{^ȻΝ,N݄ yrvd/8?'"zVC,q׍a.zJӿ82PQ3v8;d( h( e 62P`ϛJ)RaaNܰ}X>LK= 7,(XװtXP:,( JÂҗHswt~⭎|吝z痢k#u԰у^rplڐǦ Ǧ ششش!%T?{׶ƑdR]nÎlČa7uH.HnϩBSzHY'OVrMc*">6Mc4>6Mc4>6 ov5Ypyo^GgTŋCm!$ K[DV79G-nQT/ j pWCr}%~W.} Z0KTXO;3\vF~j wxwCI <&kƵ^3>r ߵq=I4?P/K%Pn˗lߛo^?ڂAw[+aoYtޞ;=/n<~\\  yI yקe64PaT3؁>p5&uckP2^,WHk+ʶƘC.1 -l'QVrݾ V rM;d&'dapphtRaܱda$uRm%Y'ltJBIOxTpB \uk哗w+7p|JtLp h [9nl-ypխ49UpActOks;K$ M'>H{Ϭ$_~==;O?՟@_Ľ@SoΨ"}䮾; oG钰Aw_M۵ܬ^-btV85ʝ>C! 2V r+˶]pgY^v5΍QJBHyߌhj ܳhf⍧f[=^H4OhEy=1W{LH> /j}׉mşxjJǙx 171f[}}} 5*稵2Ui.Ȣomʤ4룴YFMAKNپAƑF|UݱvlŢbߌr?G\V(A6[UߒD hUזMFFa4I Y@JU}I.I_(!9)1k-bm(bJEU`Ԝ(6|'dK_'{Vk^ <)uZ xI TF I݊-Šc0ɺ쨵ZCsλ$$Z3&u)o0&KbDhi{,XgB 76D4c6mUk%ёLYTZN1 늳{0aODK3R?22dYZH[iXNChRѤ CdJ y zrUFҚw-"91[ܯOT\B{ {f}Jf螠6.~חg1׏`,S̍@KDh#4f,qAg\ZB6TCZfHL%eOVಮ>`tN# (] PYjXJKI}ȡ2" + ջZ(Z0*aUcݔ6p z)ZpUTI@X$A ȆmBnE"4.aJcxVc ^Ou L+h#;6B͠j`o|ͩ~W0c)hh5PNi'o59 & #| A(rB`T@Ÿ`VʤCD8ެBP{Sm*3!(QHseoF BG, Rl m+8)m&b.5㞔ud"/k.6vy{K`OA:U KuZ4zOE# 0}IEPV/V>(2XpGW]Ƣs2YЩ ֏DE3rmE7Yr&e5D v "}X@8;׫wu2W)Kb TB%VXD(;Kb f| R;ЋI4 QB/r!PFt3 } (E?0 )hvwPJNZ%Lň@Q u0]xdWIfRh" Rꆠ%2y NH_,"E_cnXTЙ$@fJ$) A~Ѓ\@ أwG{c5EƢP~R}MD1!dD 4cC;Cє ؚKd/`0VIgY;¡k$$YxfJjGoET^ "nƂZVI JX6L[?W{-{D0dJ-6mM:A[{nzy鼾_xuEޙDӅEZ)nn@3 εﺬtFZLSp(r2(:Fm*5f՟Fj99%E.Ș4,bǽaBh8)-Pk8ݐhA/Qk CC*2xyD]qnC,&MJTAJtPX&4!JHLYPXoڪכafT( k (+TC(M@m JeDy[0zaO1'J(0RSdGf$dj"1xvL =JNmDRD:Tn:#sq֟@XV=P(-g/ULd&!Ts'k?uNϬu~5(JS>`jL1A%V$N!6 3dN ZOQA,6T* d~9pi0n3;iShX.C,2c0&D2>$Ct.Xzꢀ+9"[`o)C;oj#BR!hx)Xo; T GVvq7]\\o7~VUI:9 Ի0ۭNWМ}zeRs?'YM/匪#7R {4~GKԳ%5%Hym舔@0D8%P7WcQukIy%@A t}ZYgW񼤋C ^^q@ L;f:Cz?mپ))49G8W/oƗ[1'ͯg 'K‰Npb+X 'V8‰Npb+X 'V8‰Npb+X 'V8‰Npb+X 'V8‰Npb+X 'V8‰Npb+XNII|q~uqVoͷMsoQ} /~s-.^σjCy+H-Zp<4g]|zyVr[Be4Vtߜ)͛jNZyp{}׃a'O{ ˦}0QΫ6,9*e#ˏڳ}ƅ&SiN19wMԻ|Gul`x0 ZX8,~,eQ-jYTˢZղE,eQ-jYTˢZղE,eQ-jYTˢZղE,eQ-jYTˢZղE,eQ-jYTˢZղE,eQ-jYTˢZ>Q[둼{;yjG=lW?Ɲ޵6m,ٿҗҼ\;&zS5*$%9=C A#"(gfO7n_OJ6%Li9y vtp?6 Uq)Y`30k@{6T#bȬJ(X KOku8'WA(z9/꺾pfHR4aIjR ~¢iXcz3]wÖfd`YzD[ќ ʄϐ!*CHYei6t*HMx骱>LSЊfUR&eA&Aͽ;o9Sժf I+bg:\hzݯmnY kw\@M<ܺi6[Lľ'Z~p ޷jg9\ýٶR%%;&뫫C΢kF4 ke"w|Ň^3,Tn8^ج[2y;|:k<AKߍ6_c)lb_fvCZ;dy7*]8K..lI07s3s[D{?B kyAk@Cx0+ͫK%,ԭZo.2i ,GwZ R ܯb&v3$/}?HJ& &e5 *CA%(0xze4t}"EwUAT{nĞM ͹Mu_EJ, g޹I$RY$W(f;u}^9\w~ے] cކbƦy#(Ҵb /~jl.*LLt0wn̤K~UyW. BO}׷wF_p@.ܹә6Lrs_ڥ͇s_]_p*4aqKZ%Q*|4pŔU =wփ!qDR;XTѰPTɍɩT'T ҫLuKE4y|#AJ3 fyC߼殕Rm&b2O -`9L8Z%&J5Q q. 9 k]k KQacgaGPhI4DRems 6nٜoN3r6$a-83r?cS}o;M4$зB.LDP-3kG5͘#Y Sa5 k/Z.މz{!9h$3 Ni;y =e2Ȃ0dn{Sɴ#-<|- kN_Gf7Ow+{솖Bn(jVd*wLe Vޭ[5ooyQVk0 J/CyrP S0C'6\"\ᄥF%xg (p|Ȼ]wƙܲvy Azs1T~۟٪ >5>՛h CoUٕA\6-Fu}pm '+1u4ou9L7Z%J%%(]mI^Xl}[=UۯKk%ӝ)Qp)|eb?O?*e)&86 Sb.&~7n(bQz%;10M4BBҝoK\@3aWm/߃K QT˗zk7`tK`+Nv;\wff3ݨ-u޼}Pqv6 /-R.Da lC>>𠋩TuF@J ߄Lb8?ptVjzȱ~}rdooItVA)MJ`eQA?J:-}Lx[@G|ʎ{5y9IAO*/Gb0-\"|G 7}nv7qZۜpZ'" %i1UGuLTkql3& fǙS{fI}ouڷO潜 ^%koGnK8FFIL[)T8Gua3Uq*d>Ey _TsLGu :_6^HVnL}CdIZᔄnzp[x`ە8MlyRV,BY?˘@"ri`yDvggCz$bT3^{(0bg,8qCۜHRID<`u=u2GJG'"B*V< pB3$-]#]QXh*U@q PJZ:Cb5.1*yh:]BW@M$[:#R~\T,{ghruYK~O@Ÿ$:|9&(èwB6Pmg{wM]Lg_%Qc!8҇޾.>00ikǵ&]:VvFUꉦK*( & p%% h5izD(bmDsE %\M&Ս(9-]!]II]`⹽p)^R@KY*-]#])`CWhu hq*$m0xt(G`O5WX U@YJ`t;]-^]">j'8վZij'ifJt҃&GDWX` ]\i,thDM+)津3+9hAp< RJb+@(n:]JyKWgHWTkb 0?hV-]]qI]`M \&bV4^]GZ:'"ѣ$WFwq̃r~SCa'aV(EfP&% g*,JnXsh?BU#ۯ;l[)tՕYzD[ќgI23dRVYe:GHu*r Xw(<:cXd,qK)J)4B"-O gpEEh[q-uܝpwFx|guƸsݕCZ;dy7Z67>\ؒ aoܹ74ٜT^-.*k u~ښn;uEj'ض]עWآw)^ԊZŒW6wo^tE[*oY0I&A#<`巸,x"l?˺zE)ᆦYj$p5r)P,2X&ŵ8)wPR|߷b_:cr=].^w}^tΛUXο, R5gYagBSwy_ed p<94wn&u$7[EM?@d…Rf* FY 0?e6J%!^$3Žvh#C"\{ȁ=[S/12:S¯"%e6T2c5ަvR[R4a3z N`jR\ NKIUZsP͑ÒkPSNk>XxgBgJtd& L=pىˆα(NQP8::ڣf"ivhH:@$ѕڥaܖxR}^9\w~[j>m+flz  4-Kk[0 3 &{a*fݼ;I׫\7TʆNw?g#п з3m໿>)K?^$*DbaYwE!]O;/GͩUy{IVubw~7ai6hS:$A@{g2Hj)(-[wR Ȱֵ曰d9&}v`'SDcK$%\2N`"6);KIf6;յٜoN3r6$a-83r?NU}o;M4$зB.LDP-3kG5͘#Y5RHk ʵo-Dd4E˻Z2x񜊝t*mߜ(T{<[xP7ӷ7/#O\iHoś6\m"o"76 rٶJDJKp>KV]tt`ʙ 2; uar61_jAt;%^p#8w_jgJ;_x7ݢ]16qkWLkF+ق㳮ѕlL 6?&؉ võ?wՍ6 L7J(t%;Jκ:饱oW P7˫zf& 5szZޜ}uD-eyy=NA:TE\(_>AnMe/_u|8Ywu\;]-?75ޞZ_|jp38 2d]QWkև899ɯ7cqiL !{z.yTzfy=縯V+jxskVavf}mMkhsehm+wy0B4\aJр wy7J?5tͳhLВ\A"*FW%KhrǬ CV%]/0r3*f5Ŕsv5I]y+cbno>XpRͺ UA֢]1w֍YG zՋѕjLhx]u?< W<F;(t:Jͺ:饔˂t`"YRtŴƮ+P͞u5])I΍_W D1]1cSu5A]i ]؊PWRtŴfS:9j/)HWVB+U]1-)u5A]/R;q.EWLF?d@&+P ]1btŸCӉVJ7v]1)ʩ@E1d\J3v]9V҅t`%˹θڗ+ Ϯ@g]MRWA[! ]1-GWyLR1Q /GWe3s?w +?2tuFt;JϺ:)Q _+2t%5v]1 &+E4[QRthkƮ+vu# )f2ȴO}x]1eu5E]Q©t`)M1b\EiuŔ4O'+)g;NSF#no¦0;+<&oBV!̭\ʾ'dDƕT5;/[R#M(hFXV J_ʌi?aJ'g48\Arno1qi@uLI Ws7izuŔc[eճ[ʮLϰ__b\[- 4fή&`.n(cc]t~q%6tQD:һoOۻ^ܛ| jP>#k-޽P'ˣvY7?|ŋeX]䏅H ⒿZߡp7vAltVky̠їmъ[uL_]8*hc>Fv.:t_vvwEzV܋MiB?/vYlva9:j #+j8aQ?i]1gDW^k(Kb5R7MTUEz﯐R틥1&Օ5FH)ժ1 Ɏɒb }-b1tf4cf䘍I6ntT2:(Cu#h`@{ys=EC@P)!dUK[h4N;%4)hR5CdU5h*F)#z_F2:M]B"92[_  .XD!D@x4 -сҾ7W1O,1IFIdC(`PEc}UQH6k"OKb6^ɽ*I'rՐUuL17 ω$EYE 4BrMUTB&:Q2>}4 j7(c}&E1cW#Z)| Mqm(LV VÐ Q+Q.U&:+T:¢?I+ub)iOv˄:$%ap1(lb]1jr"acV uAZRFaQ=f)xѥdq;0a[d4.! z7;)', c@'Dc7B2[Y\ #жvV7Ю0D)Rk,jlekJC2(FWZ8l#S^m!-C+|ԎTS>J>f^Xa~cb4릪GEH>yws 5J# (6Q )mPOXI`$A dW|#/لiBW﵁o',dy/1곐Q6BUeZk]#ka#Zـ$zyUP>;1&Z1?9;om3uٻ6l+qR/9 1&KnId;Yդ(JeQ%aXdsUծ׮iV/;t+^K+fHNV/*08G!N^}0[w&4i|1_YS5p>%~ת&'z|e LZǡmZZI|tx ŀKyY=T~:2UWBҥ`r\Jj WQ8dgvڢR{:".Q(JR|$L䴬yE,C`rQR05`zOA `}QHVǃVH Ef06}`Qٽ*YȩՏ/XA}(yYQ‚dNI.+I!Hb'#ҷKӭ>}wy2u1q׆T'X.B'G A]rk">3R7PKmnQ@R@"dePEx%tMކVAhGcU;fCdCjxrI'#66^5H#lܶQi<8ӱg.@Gh⬮K,@$ frP%VJI>@2~ȃ!2ءvGycP)EF WE!^es,0 GEZb+'R% fcv4] ^{Y ̢jTRpP*Fi*U9z-ދUV^!-Iρ$D}f|t,2zaHuMmZ4+kt i>jG$2|45Z]g P]BV  Ihd*"ltLowG(8w`)jY:Z 5$QhS@q8RV,4z;fc\p=Uf6k80)QC^"oZHs1#J -E{r).2TA(< U@%'' HvXoYúI!>G/[`E^H"\x>\N.v)&Dyư2aNXPQˢT(:#>L =YT'X:!aQ0'#j hNiu)fndjRW 5jf=_8mOR˃>T PMIA1Xg]rmn.zMgA*fQW>Zeǔp(AJf:t (AB~)AH85]8 8e݆sAg[1xx@A4jk̆bتn  RLȎYЬƾ7"SJ,豈)`4&*? tWW$"X0B= R GV^ōbV7 )FŭjG:]WV'yN%!(Y4:mD/0x[דXnLSA M)uGϦweNfd1Z|:Yzw=97g:[,S. >ͧ 93-]٨4XHP@GpPu|<=meT0]['A-3EWu\#O*\:S8W 稊irG|#&r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9]'>9=ph;3 dXjDN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"': g9(':B8t'PJ tN Y/ DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@:^'\>98qpuN @$9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@|k=[uIjZ[m/W?Nr M/&'@_۪_;Vq (=(Kޢ䵦Gte1 ]UF*Z׎\-u d W?xJ-U$>5w WUEݡUEbNWCoNeƿ<]뗦]r^vB+} @WCJ>snzCW}VC #+aGtU*\BW᫫r;UXpp ]骢ܨQW{*گ4t(-1T}}+~m\J'RkEtuteJW ])vte#:BrIzDW ].&}+骢~tX컪peo]U72dꛡ+'-^3 `+K/96KW;j'ڕ݁,cs&]7tU*h:t(%:B\ \C+*Z)*J剮$CzCWM(m_JrR c+%83}*kp ]`骢$:Nr롿Xu%}0<;NbOQ;v8? *_ݠmOt;BPToSs5,o|1JucJzE1{7f(LfvYvp]u+q4fOTn07:?o[^MMW{\T?t[T}(hSjoiE5Q6IF \V#FFIojs"Sڟ4vC AN#[c})elT!`އ6'}|OyGT4n`*Z(=]?Ɗd`Е}+S#档++蕺`#\D_J_IJai1ҕ䷢Gtc?W+\ۛC@_(#:B|W]fx}]Uy< ]=ޝ0?]^j {!'N(;+]9=WNRN5ӽ WUEUEi%ҕ>rJWp ]U:]?g{]= ]I= ؙ:BWNWt`m*ܗ2Z̡UE)hJ37/ waj8٧.sXVitv}aή}@ݑQ~?bsy3qy:t7{W.~XoW16_~x= xGҝpۋpۀGtv׮w6>gM_Wts|νIϬ|a8\L) X*|>M__wլ|lDC>w:6؝lf~NF9?yirێ&9&_x|htV6Xiz1]e0Ege2X}Ո kN*d0itrs.&/rl~L?L6& .n6Vm~2][/g*Re[o3fٖm^z t|_{_ IN/t^|#М(U >& E0+]\KG_/!v5 |c޾ZޟF귯Jq K2& <l qUJVS/UxM/s6'Zg,Qil1GX 5C3̩| вQSY .Y®ѝʻK3τS_ sP?q;|/WZX&WT7ΣTi1Ul1 3IH)cP ж=N|_i6j`"yRƯkM1՛ 筱VXERF陬h7E&liY湺.RhFd6kUr $*sKW!-(LM:oI X]偛 YWr/43Pe6^N:& -T@R@}qFd\$AzfYz٪,V;lv/]e{+-E:]|S`_*Fsƍ]o]_\H|-H~ԇ0$SATsmVNyƍ;+L~J# gT}w I[`כJ5J*L~5#|Qmބ15=luPɛ&Q&(g4EI=Eיd{gyMη1&M$yZ[Ϭs)- µ6ƜeE 뵏f#\HC>ZmUQJ';R+b+L78_qiy.Mygȍwg_Y hmJ; yccZ]\\ N6u';ㄏm2mjniD5Ӕg$x3ƲΙ 9U9{DIq@{2? w^4mQ' \3\n7Vd'i~3X[-ɲ6)x" ΐCrW{p[BDn[Sn>[z< u֒`#GVL,Mڅ a!a^)NmYrAc\x$.vSJ$_n.&i{p*_YR<~tvKrsw "WɬO&Y*hS?" :I✔^8sLz(Rp_f=$PyA;WCPoĝ͸J=Dbi-y6L TNdBf DzDpzPr/ z\-ry_b>4OR#mJDF%TpɕSDJ]ub=f00 l'揷6TCJ;| Re3 >( wKbhQ\:1Ӧ#m#-[q~XfR.Z^8QI\ʓsWe΅F[Nv'-'[ XxΏȇ0OENV DfYvagayh~eI1T\$_{=OS'}t+!&l5aPNXOZrra&/h ?(ʼMbU"'ڄp,((_@ mQL>\ިS4zS1Ww1|@'+xWR1-)ǣΰWy *a0egf2\Xq CH\\n~h|b^\.N0su.Kr;le jT)^T{.YR8uaMB.VS53iGQ8Η:x҅2rwF.uI[eZ8&=+`>?!W1m(/X BĻv8%7o:xuWſ^ū `O/~|7 Q N߬^۪T͍XjV~o&/wP^"mZt Go>/C>&?43!*nӅtKt ݼ[f@Q6*=T@iA!<܀e¯p5nIbTc;c]ugCg*γ,Sd8Yy8n@ 6/e cf"a?' $ʔ0!^/YvAK6͂*dB0>fAr+'Ah&ɈVN)t|&gG&]}!/a8PZTe B);3."醋Pm$ɔc*?XBRA/FTTSFwްfj+P0?|[Nl&nLE Ḽ.L'1-oM9UVom5eA J+Kn5sMnO=n4I1[KaɹU,ΘX"$HadМ&9Osi&NAT%hlf K-hqb k> ,|vˎ'gb;bT~y z(v)EI.DOp;涎ADkt@ NCsD8LFt+jsU:l P2":@2'y|{t9]!\BW}+Di~|EWײ5;]!ᯈ슪--e\ ]xj-(\%۳`ЮAWMUO Ԉ6tp`e`Kthw5վJEUCWHWLBft5y;ݼW}\Я>5wbELaZ?.^XIMHXƘ"[DRH0$ΜE߾ ?Dl%ۅx\mߍ.BO2 oʞ׳MI\e el%6hr<)~yϢZ~:5 >e,pɃ]l9^&xIu{T+]b~pk՜zaHSɣBG )V)Xj ./hЊ~Α00=BBT "Kە)$z4>,3Z-o1}@7( 1|֊^>9t+*KV?1dM>I@rł4ZÜ '%u,8+TB !G~'(o:7t>T ?DT8+tf@gx'HQ1A(8p#,7GGy4&C*:thvՆ ~ vqCK n)3'Nhga2Mb4Z8~ߖۺ}?Оf3sS48FjMhүoɊTs>_*M[elBn?Z4H_- ɩ]!Hx~0z_/l6d:DޫTRD_-k(cno4.q^(3eL imz1A%ȶh{f7-Y1RSDJTGfۛى!C5]hɜr",.ޞܷGjNӏ|kj%kԂ[#Gn.~IeRWRޞM6XU'lM75< ъ~B6Oĭ ֈ0 +u] *t(Myt%z6}yy}DhziNyC(SxeJPc|A ;YǞox5jv1 @qu(BʙU).~lO[W%gXbW`ZD@sk$>\JQknCȩWܜ(D4ȺD4S !F4E ztC^x(*G.u ;C#RlIm>Np(օC-eoo8p8T;JWX1QBBԅ&NWϹá+C,u+s@ յYs=]!怔+Ku:~V{OWR4j_ ]A`%i0!-]?ZxwBW}+IV݆6Q=eLiS#k%kCW2UBW}+@,m ~yکb ]1fwh5{1KxCWHW3) "`KjCW\[w(+}4tu8t%4ZֈokCWWѵkT؆ ^1SJAua:Nv4"T񄧞, Υg;%143B(Z=]`_͡M̎??:CYN[z~TtxDw^۽^]F7 K 3γ,dP +JU7n2~iCSzbeJ@k˞m Y|u9 b#{FNlp%4*{#Jь8r]IjDWX ]BZօb QfJ sFt%"5µY ]"]ɪD?]`f \NE] t(lJfׇx~]\%k3wh++eMNCWtEF[u zpn֐d}U56tꩢJք׆.u+@kwB2ՓS U#\U.thw  ]]qew ]!\BWV3t(oJhf] S++T] >v Qf (]9t/YYauiv8՘Ws/eM+7ZӬ\h61Hk('ҺD4V}he&9FiMyEEW>CҽA/>D҆q.kDWF[גd%# ] ]#œ?a5 "j Qƻ:DPE:%GWHWXBk h ]}=tVT=k/B[ƷZh-Q0JgSl b ]mz`sSAz&.}y?rvv9|{e4 (by~N?Se&r:vσ>v^)^ m%( ;;!g޹shu21Q^3t֓Q1E]F\@cA9IE̼$!'TՂ>?Z /UŲD ?&G{ Ǐ򿃇Vǒ:jA: ;vmeG(m<qI5mlwl2SRn,x{ }?i-pیM}fKیhwQ*Ӹ͇6[*K. *h,~Oj/NեGo?Iי>(g;9sPA}R3*2: %& Aݣ^JUmB~u0LmhX5'dbH7j-z>{0ҋѽw`U_!sk 6F^,"mיk \9tn0B[mqnpZ=pzSQ '[r\㧐\GnM.p}iM񀴰٨L8~kiѨ?L'/sJX(Zu1NC8>KbOKpp~ZA;/@w]oGW!#!`] Ogj(sH!U-4/yIJQwuw=~U]SVXq 6;&OyTF/X^| ݣ8O l 0EHe YcI"mCmľ 7${DJGkhNQ#T!"CTDH9嘖zO H͵B"9NiQmGsT bjr؂樂Z )$ Z#qH A%=μ? Azc{X- {BA`}}DKmA+N[܇: mc|2|T|XNʨvi֣.D3ux*e+RaYg|8M0ۂޣƣO@6{4,/JbkWaafO>+RViȞW~1딌0gϧZ~g:6LJ.z%4R mJ&@A@LT&62r%=}Ǘm]Saտf_^t *$}z+&;o -ؖ '*f[w0,"$)%hb%J׻sF}G ZBKT!< dt}[%8t$WDfFj H8F86cG iXePTQ鰡&4EIp'g;+S'֫d#GG2Ԋ' O~RV̠h$00rIKLMTR#*#+TT!H% M[i(g\ĬdC CUHPA?" s9`FxpE@8a N/>0s⤦GǔO 0qܛ1ZX@ZKAI':'FWT`u34#^"o~HƂPz S("'[qs3{n<&aZ\?W}R LW}C8o+8nY]i:e\ @VjNE4y.Dw#"@d 8kE)|Hmt?h"{M0(#| ] as5 wwV!c27;hC+ebtR$#YH6}doMt6.Mm+Dm%$t mIh3qJE2`NTI2}*~R{Df]oEbrb|SQ/v'qF߶щ9I`#eш9E[oxO~jʶ!>x'" 7)E2z&I fp$5(WI(sڈ#XS&uap)1 A&c*5(K)"h R59mx0gϏf}_ëcYL>~b!*e%K qQ[k 1D@,A4.XqRI` 8 K%RL*{/e00 KzKkPָ狲mL?՟W.O`cɥ${I7G[s'ӅIB +pm"a-:RLݙ)pcwgYl(85iatH\)`CrZ2Q˶;Фaպ&iwkJU2%׳K^\onWf\1XH]&&iaTl2 &`+ϠTlE~譬htz,u`dɇnbMhti=\zRtzpoXsҠ?0W\_~|~~ͻ/߀v`\HPtkz Mon>fmM ͦ5Z?:/FyK^:}7Q/~|7˚)RfWO+Wln_tS%ZIwKA<Ņߥ?@Z1Vjx-jxca=锃&e8KT/ucy4wR?GN=o۩#^}   !(e1pPJE9u{U:-pR3"$D]^Ĕ ^Ͻ̥2(ywNKz5x!<*Sqr7`.=0T vֿC1Ys2T(3-c`ήGqB")np0Ynu '1beP]mZ]3rnQ-c='0ttVt MLr] qp zƌƁZ 1b"i8R \ʬ L&1O[?_/M654\2_&PB^oaKn뎾:_K^iwq 줈y]m1MC0ﳔ_Cf<[ 8g%X#i۶ܡށ8 a2‘O2ƣA;kڅT" K9Afx^k!m8v8CuP.v$SeptjɌt>~Дv@J$vV|Oϣ?TQ*cRb2叫6P2wd>@GCP* OXbRR *ˌ6"%xg2D')J/p8% %0l\}[#KFЉ̞"Ay3\HGjf]d̕sU`w/| 3dP-Nsy#W2R94U;mEx%؝OWSlk.?HTSjlbW< o0 d :*^6JP+eR1EYR&zGs2Vi['⺺(rEnPv]Py EOĂƶCZBZ\ tBeRLDƧR&X{D%]oL*Dc8UA{ɽ~$gWw;`*]10=TaNFtjQcW@`]-oB~ :?Ʃ"&Xޥ/_^Yb\Dz_36QAzCzV$2\"їHrnfbHB*|.b0bXJCυRhͨL/ERY-}r/ñ .eۉ.zcM9r~r}7.,Ťe﷞A ]7?We~jih^27]òw~ <#ya %Cb9KbEZqaL՗J{AgT_dŵۜD|8nZޞ ) AJ_2XҋMohSmujF.^e#>v8Yے״xf,ʜIŶZ$b'n[ AJe}ƃb$] Ŝ8h8)uqY2۱b{e acMsSPFI'7I'?e0uQxWNfvt1ERe֤PcmNRxt|}jXFfֲa1`QV?ͬɲ.8i2'-6쾿SYny v 3MȌQ2@<Ȱfy㭥i2+̃tW g3}[* [YNE;a1n~ F$f wV[N8ʖ^Lҁ4 QG՗H/ 7df0~GQ,nm< =;.[dhC{n7gf2p,g4 5sP; { _dir科HzfPq =7̛G@B/%p% Rĝ惟?wΧ8Ӑ|8H?XH-ƢKwtm?7oZƋF?Qb,]ew}S9ٞmhgUe]b۷g9Sz Gg0Dv#}msq2OT7ρx }q<!2b3ukOEaײ7MH{D^I>-NhtPF'FKjg,;Hq*!r=Sipۂ'']鳣=GG"/ĢH}V5Hj<fW3Ig9Hnx8K|o߷YXNj5H/7^wBmx_zXv+z ;d/ĭn9oJچb-ųmޔչ)u]u]W+]JuۦVX u]Tuu]麮t]W+]וJuN82@0d&rZx2ZLmr){+SITDuMKIT$k9p T iAGٻ6dW d"$B`B!Ԓ!Ufȑ8#Qr6- kZqXt_P)np0sH[]D‰uXTF]/ ugy瀵L(kΛJ\ bZ1b"iZ>cn]{0kVp.6bۊC'zk}C0Z.i2a^t2ĩꁢ{ZM:_pX3[3cQd<)VYjʈhA XGh#2&"}ogԊbBd'Ntĝ?rHuJc]ӘG"t<6ʌ ;6`#*$;A (Att!6 wh=[o>o u8xUԆm;Ӗb-Re6f pѶcAٓeRH0/VJe`nYO^= Ao=Mzϼnbbeo/確zk@l^|5 = F`I6Mj"{1?Pzsz( Rkp,I)1Mcyjy6C#MJT='%Z }АGF|<ea=Vz)/O3z\T]]́X[`gEqIKGozt.uw>|>{PU%O¨Ұ 5V _5:2ԯ8tx.Xy2fknXPO*xy }5€>^fY4iNSLyQ:}q8zgb*@zbdց(WZ8᩵ oYj<۩ AX1i!|KMTR#*#+TT!H% M;ֳig^Ԭ6AۏKmT*h(`Rkjj9_iQDf0*Х |x)?7-!H<'NjqD0zLydT8b%a)juğ8i[@ Gٛή=~/4= [giFDނԑ*m85pq6i\i6s;`QdbsTwy̫sŷލ[Ejc@WjntథA jW6kIJͩ& ϥH<v3 ߇fz(י)I X>WWU Pz[3}wAZpp5|ڏapz_"?}~a<+hRLWI@(K lp` JSX r bLyuR%bI` .$ 6S 9?Vb-_s-1`1CZR%TTEZT3<o~vP%Tv{3 ۨ 1t䏏 O~s%W>|=/.}:>L7(QmpEn@HVDG#>E)kJz\V|ɉ|p3*t.+'ߖ.0!$E MnbJi5H" gK$X3i8x +g$9gmđp)ap)1 A:1jveR4zqy|OWٔCZDmaX6Z47|D1pHR%%4HdùcF:g !jDsfu t&pp7S@·rCBu;"E RrȔ Sfi%1"Ay٩A}(İ0//KY|*;%^q TqύŌr*F-眚OGp9{pB,O"O#\SKSawa4p# d08Z'(EΊԭ"=d51XM#\Oc'X7g.FZscIYe^j -ƃ0?U!I[2kX.CIƅY?՟Ww']wɮKCĺ}yoy{NBǛwo\fAJcXL t%s= j~U 7oTIVv]҄(@M_A@Z1HX&NZe` htf8ۈncq蔍&ceY_\ӦFf gexPQC{n;''fzx|?x!1h,l6\Ef;CQ*IN1$B鉋<u>;t1Cr;``fH#!3DI)A {+X;H^dvt&/dC}Ny50g>M/zޮuwB-vgA[6J!^ؠLV3Ń4'`Xί/ Ï&ǵYNj5e6Kv.섲{HG{U8_5L[Qʤj&D ? n2:.w 1ӚlZG;I[c\*NE H ^ wE%K@)*1Nꌜ8A:Ŀ=ztQ^?V ni:~^>NӚߝ?&=Xzn^]&}o\0-F9[WfkCw.1'BGB1NGGYia./刟w9-\0la%iC D xlH$ ) Cpc9GM"Dd)ZiSHD9DŽRn&x9Ȱ, Ƹ#aREbӔ%RHD#uFnHt $Dŋ)Hl;IHj ƚ`u> w<QT'hz\HZLW]l p S\YA",z[ |{GhQEpSb@A{ ;̾_f #3i0œ)K(m$ٴHyӏ /K2Ahٵp#ʧ !&y砲e32̀UAG{2(:o!FjIDk1"ik45[8Dl̰nuyou~EUYkngtfsW'nzVXn. X%r+iӱe]qmtjBM bystO.7Mw4|fwa,' xrq@vfN˫w7>/Gע'd<wn#S[5[TUnK\Ѷ[0]dy(:9G]DsO9+H^QxN^:yY'/e *s@ٜ1%1HT  PlC@׼#y{^׼5y{^?`LykDyk=yky@%FApeX aAui~oe,:uq/Yb`\b҅Vpb#V>h$qFz"zw+=\(wp)BSϘ8P=t:FY^16Iw^@S]׍ʺyX)n.}im\a4sQ~CYh060}uO_GQCD͓dw%S,TON#(5;ɦXe<,Ez-nζɭ>C'z }e+ASYɮ[K{A:x6HL]d ԉX%? +,}Ir~wؗu qM Tk%hqH OkrdBJI*Ex˷7|fa,J n0B&ƇC֖ eV1 D1%bk:'Irйts@>Ѯ' 7W,tnl]叵8X*7Y-ZG]™h%O#3):5cw( Wg|4*5{v _>\W׻MP0A%ybxVe~|4^*+jJTb-!GDv#ZŒ/Ʌ*s#ssVGR,(ac!TDU^2 75aV7x:٠<5F50/Vnyn皪 AFleb%~?\5J㛆z9iPnP$ySz{1ԽÉg*Z Ae*Oke2-iW(|v;K^51P5c;#ՇLoFW׷qN!$I΂RƄ)Cي`H `RNǠR)\B)E֞qhkC8眕$U7D~d}W/vFT'fFҠW"hfN-c6 D@:tR@>! KRiH*MEcJFa bPhkQC Fj9X#cOHƬ(-ARpjҔ-Xʂ#j Lh4S됱55H.%%lBQ#AlymMveW|˖_Wt%N۟Go^n|ѯ{]9|16M,z`$1HPR+(%y\AZc?4qy~ >~VOeB5};TZf5Byxfs;gs{1 :;)A&C5p22qe\yZ6ͥPN1ìdF*oy4:m*e6uz%pֶ b$ZipFe坱y9}غ_f9L.7~5J+o2ɸkF~ʜkUVlj\-EE.Ze<*8|dc]<ƃ'[:`$qU޶ߟ͹ #tY g0pc1,WR %Q:ZmВzj) 09-N6YOɣ p;j!d& SA cϞVG[C>ufƂ {)mhPX\krjW*&)8e /E8_'߻ZSowULS6޿NL9/U&p}?dPkLBwony$ QM> rx4Zw=Jє|k_Q|^tVM|@rdd[ƨ3,:Jy%MB8 B*u)B7͓B|HobЇq鬸4j=-a( T@ T@EŤA OB "$pe(YPɂH.D#`bN8ipێJܶ3!ZFbІE7+ƿB go3ɲ>w+G_>-%r[y9%ñ'9ᫍE)|HuߤD< Qf3L&$QAu4 hg^Hcckj ?t!Ztoqt<̆\w}Y|(_)ehgTqf*5MHh0ļ&cB'L/d88kM (Y@4HBViQ$J?UlPjo (_ʹ]to"Gt A(]t(%NmYHn('`$\z+`KH e+L3 3xeTj4c)%AEu>L[=YKh|j(8MRrs ?ƣQ'sRp9cSJ` Zf*?M\`ZE 6L'>X!(JDt(Qq%@ uwGAcPȸg7/! $xn>iTpɕ(IK)K\>0E ?}c LZ.h.Qz`8š)d@AWY^dRn)N]N7~J1FV f  08yO,ZOtSD%d;_0t8-Q[t>)wJ877յe8ZKdqocj0SQB9aǗL+)c7 r<8][LM6D'XO*/5mjrtFnҁv@;kg>Klv~,\1~i]RLf gI{.:==]\,~THŴhOFa%p sjfF2ZxTxqz3AȚ0f\SuݰEVybb_ꨌKB2EgtPN5c[v~J-Aꔍfe8=y:n@y vsƾO3D @ ̈́mk7nxk ) )Z|>h abS OqFOrJ$n#S MwHy2XrLAc X1'9ÚCZUrm~:O\糣**~ ߏNE?z//;']㦏?̹ۑ9?yBi2vT9Z<2ExYr2ĖZ^#gl21J:qÃIEA6#jA] * z=oh9]MlZ~l^ =R7L|\wG巟U5~:ysw|M;.i|<)?6)vz]oƵW~hp=AQ4Am4Pi+)GR%Y/[e`mGyΙ|nPo j)]O\9L0-H;w:g>d,E2Y"Xd,ceXdY2Y"Xd,ce,E2Y"Xd,ce,m,ce,E2Y"˽q5HLXd,ce,E2Y"Xdya"Xdyg,E2Y"Xd,ce,E2Y"Xd,c.ce,E2Y"V{U.&˥9XgpVjHjRg^ *w\|y78A%|Vqs z JB .`VeLJ>Fc3jU3B6}yiG-d0j9>~QBм+!T\Mi2X"и#/FCۡ.l|BH SRnfkdQ]d@"Iu iRXЬC=w#yD4J`?f? ΅sT|K̯e? g, F`ŕbMި0Zl lrt!H]Z9GV,c$\7=`SK6ej>pYZuhg!pնK~ %m~THmggss05k;Y%/j9AY]ռJuDGҨ q(h?u^*t@Bב?}&V{Guнf?#] E"ݲIzud56#O g(f7Ta@9rIf28L[oaOx0Ɲsm7Ya%L.׿ vMl Yb!FZBq8Fہgfߎu(g09߈Tdw} vG0WȿZb; < *EÚ@}yM¾ ` #QHY!x z@c&ye4zl5ͭHGnx*tL!D#/pKlr_AEm>{!~qvhUUJi#ͣ^?`@ |-gZ^jL>vsO|Y5bsnҡD$n$ NN^ 16RE}&݃.Nn>jaw7-t0w\ |·yS[|={k,О+`Ev2ޅ/ ?w`-fͭbm6w޳{$lMwrM.Ԭ\_m/q\ aV',H#:X[,"L0mWr=9_O婃ؙS<2,|ՒEyXQdMO-%]~8a׸QZ"HIk\}>L_.&K?(iTݟ?OW#X`T9F$㑅H>HԔ1тFQ4`pHnԷ&.s+~IMys&7WtJPm o5:Xl)M+̒-zU ,Rg2D'(E4^JJ8%Zb%0liCo ue،NN6@%ػf*ﶼ EWU5sff̊y:\ˬnѫ+VVVڿcc2^mJ$v< "޶ tz~!e+Z <[u]sb{WgK-Mw0ơ햄X^_ӚaN̎ߴ7\6 [潢MQz@* "`=M}=]@_8TZr։6}=7sni'nЪ (o&I w?؄n^z:뼑)R?zhៗ~NnL㒰 ~nj T]A `m0%k4t~j{kf-_5mpmd;Tن{xOGշۋa<TsupS]n'6|nP'A7{6?.08NLw`' ՟~Xѥ!|+˜RYla bRc6(SX8S_z-/DAmv99Q裰ǗJ `@#`Rqyη GQXp0ŅAq,BbC{`yA-z /̂vt>!2|\(M6ׯ0J/ވ F*OThi~h.J`M  b:Pp.KX1܌Xt^r-" `BK 1Nnn waiH,|I 5GLe8,EkBKaзW݆h | `%̿O0 V L|2Nײ*ǟe/oӢFڢý)G]u.&RV jOUnɍ,Ot9:QHiYH"w~-R!rFf U`ha)16_AsCܿu-bb؀!(UXGDi))U-x!ڈXV+;rbr&NwMm0klpK7ҝ͗BY 0#Hf 2z4מj'1r *&24(Goor6U_r˷wm wD+Ά67'ohj~DOȕ/Z%*lO~F'X/¬5qm쥳IQA)3ZXDc2bƲƉVW.zvy l-sF-;ۜ(1!Y ;BeUV돊:-EǸ..h!10 .f`rH8+QכO\Rb- KZGK!W1ȘpV{tHbMZ+IϱV;¬0 } c*+%ZRPG Vp).}Ԗ@U`̙[qYpmزXpS;yz3 3?Y5V)|4 j-$* %^Fґ0хTm؀0ہeߐE.qVq LgP  r_ JAN" M+K;J,YDz! #$$5pFBb%#1))DXҠp0I̥66tlK2kZִi]jZ5X\<$Bp]YU/JE* xsgi:(UA$9 dL"W@$j|1j2TC葹J+sպ/ Dvsd4hr{)uZ:eۗyT/<E?g΍~[Z%|5^ onMb̳Q]oo>()Rg)!C!wdYJvqg9\G MFż/Mמ$*;hHJ{d/Ph\ bm";`*̟Kz7=EծoU#zqֵ;e num򺖒J/n8Ogfh286\0(3YxM4u~GЌ52wO\n6۬U|>~|3Y#j|34f*gRe6ra!-X\.gZA-A򤑦Yky̙#p0.W'GU(de"Ǻ7o='g_[NOkV?w˻.ѹ_وʝ'.;V߯g/3Ot-R0odo<Թ!tT21 s )K$5%lBe[#Ҏl֭̍΢!un;xX.Aa(Gk2>4e(f\rAiM7Ǖb }EaBr i+z uH14<$AQ.6!6BkE8UގsDEa䧮?P^o[;<xaVm`;8|+pJL E&1arUc"^C<` l@+e_%$0ͽ@Q#1HbW(kc?D IQFFXZkiјJ(&)MyЂLH.8"Gu ͭ0&hiRZKi-Κ@BChoKժ: i_ɍ fH7Nʳ>xO%W8rgA yZ%W3J1FPrqIuqI#xKJrtٿ ln]ʓ*ޙTx$ (Pvd"P$"dm,h<%@I74 8\p3Gy/یX)IC-KFTvé zF*($I˽uV)u"V;J$j'ht F#9h\g1XvR8aTo85b2Lt&| /݄W-e@@U2ŸnqYרJ#~YW^Ō3u.z]Wʊp,'@%a EŰ @='>"F'̹2Ra8p-6Fj5ì: n24H]Œ:|}2>%P'ijցljY1'lS4{0J('\\k9ŽF.FcdggHq+yDgc,'zpWe^*Q3Ϗ:ӏ/sgKV2=% }~+uG^`I0Bdnqwn>]?}Շ>]!W~Rxss0{d~F447[44o]šrźOȇL/ ڮ%q}ws}Ӈ5hy/2nV2N d~;uf}mJ 06,PM o_"phclUM`UgT1H#H~:夙ng(Up g2B<^JgЎn~ =mԱ5!ECK'4YLᨷDmhVV):0dkԽwdƟO0Mh!9xe|bK2a1XLVDbA S;֝^B5r uRʍ+W*cV6mq⭸ GӸU2Js@+KpNĂҤi;C+?dCu(o~JBAd1)H$Z)${ҜrT4AFuă=28Fz2#@'x圹NTnz*O\Ӆ9A X*ŝ͊_`t`v C ]?.&7Omq䍧YN# 9SMdOטyaf|}{7s!q>L0ǃRL DL &ႩDQ -H! Xiu}]g twٸ./~9EQ9DyX;\z ?_O)/k馥S.Xf Zs&8S2GR9KVPYt|K sYO0SN~pjMVQnxy'KtY-e̖0[<aX-V %&uu0M,*PjBdQ gԦ,> I&x\'q-;%.o9n1#B7 ӟpCq'::0LyfXxs`c~}K`h _$^fNi,%`Q& k0=D!y1H L܃) : fˌBJ-jaAC0B"5Rj%xL,k+*t,973k,Y z~Ht $2Dt/r; ((~a(h$XϖC4/Ə S=pOE$FqϒjDԚbBJ$+OgNV\`{ >쭭^t`HM7#)dU\ DQҬl9#.@"5R2eLr+gs<[f89 Rt^8NO6xxBCL5c϶=Π J%α#G@P{G@EG tTUP!QP{G@P{ *IP*kt}FK"hJC$MLFiedT!>kV3.yε 7wĮ'[g$vbc'4, l7>_GS_-MEY5{m2r2*="9/Hid,61п~; <Ζy0+Ν*sv{=KFJjL,h6ݚ˂{"֑*U[uT)653Q٫G:< AG_>].c'lsYKSA^[]|c9}w\Q#-EV/7|Eܕ'iy"wsZ3i~n䞏ϴ>E(FRifn1ӤXY ];j׽ q\"AI[B'+<4+=hO_H-O|HϧK**7>ߌ:=}wiWZ:v¿1i6}\юkށp:㷱>!$JFD'EApz<:,LFUJMn(2DIM%"*X!5?M)X\{7U_o3S-u- U=r?lN)\ʒ $.:mS&ϓ<&FUj>>f2Uz%9좸 [ONbqOhhty4Ϊd:IXO/HcQd$ 3Y(@&G,ZׇV?PgWdaYcp+4DKƔE#Phhł&D±hn%U{E63f"ZIX.U"C )DРVଠ[QacijiҪ==Qú=go]ԽQ^$uwM+_V۩R 7ƐxLCPڥB`ggeF8xTZ!Ֆ RgT^O'~ =@)n3QeLXA *#v(''˾@1lQr˅k:ztA Vv襣n<&9^ j/owO_:O5 q%*x 6'?$AfDκlGiWA:/ѥjӴL56\Ǭ7 bE:kmohBʘ|1Κ/ DL0Bf24ʃi'sB)[ב'^Hifk,jW7PTm4YRyuHkKHhij!-2(k|hJ449u錪li At"&MTI`UPm{#HQQ꘼8fW)8I%$Ιker 7Z8+sXu[^$=']i.)1AIuY7׎ [ *`PoKPj1ƻ"0i:IwoN8=N*ϔ>5 &5R }D:oKG&q˔o$ߎ<$ە;/ГT=N:JȀZ񬒗o 63!joF9 28Rva`6hJqR!q(%B@1 &A+pM ȇL-y'^Xh~flVƆݸʯR)~q)(!Ia dwhکLڤdg*!3Q ͽM]0:Jޓ xH-$N hi¶輏i<": Z%mek.njv1l=MEi|H}zyN Й~.kόpޡE''""b0BU<.XWӆXG_X=[IЋiXu4)M{J$k=c/K; Z Q 6n;q+!/6Y);zprxǞ P@@&ZnX(3ל((@]/I&{5oǚlJ3eц0ץX $`T1)"1Ԅ$ % AT;MYKzX,y_lRw)akguAF[QCdU.X))ЧBh3Qyb7xAM]LKϾN䒽>KZ樬,8YT^eJGpQdh.`h'{= E{cV'gnM?leHNCb I`̃ZmH2M9B'0[QqRSF16K]eD}2 :fQ09M.qNF^ _CzLwx-)0q Dgw Ν-p1yϕ=%'rvkf wy >~*'s~W0vey>eԊ[YēL}'KlZM(()&8㒉svpl\?i}= ű0t+GG0x,M:\/.h/M_i@=Ϲx9:?ic?o]bhd_^޽;~h#4FӃIt=tL +NJ7jFُkO8{0?6]?\N_gkq]2d>Fv%QD6bndǗC;;Qy'66vfUU>&>Fbƣˁ^sx:pdDw.Ϫz]e<.-Xl˥/d7bVqQC:ņA:ן??X~?~pD_}uD#0X"Dp"iu뿿e[t孥Sԋ1#wpm@;3kk@R~rqp s2hSLtw=Nty8%ZvS*uJOSB!e5n\έwHr0՜c,9ZV̻RJ4& 0HR 7l/AGRQ*f_B (PHN!FJk I@YXU/ Mb@KJ$wΥbM+tQCȹ]B )8f+t~3H5c~?1%> ڼ 癠 R>_LSkާW*Y x%:xv󨞢:>G16Ql@jQ QD!RR3ϕ=k@=J-ҽII@koQdqf NP|f'E!: %BNV% GV+7gMr<3 -qj!^"";BL=u1Ebl"ՀIΩA.Xv wҽCt؂ޡųw7uj}k^ #Tf䯾_iDCQnzAmo >x0d`pɭ㿽o{wxK JuVnd*-i-d/ꖐ6z)ƭT7A*|T%JN o\Hzn{{/^䛗]T97i=B-֪itX`g-QV{uya( 7G9ua!pdSH7oB6AEKG] u]_X/pu3QҘrZJU$}L,ӥH/Aݖ+Z`+EIq VzPt[ȬכKޝma3A=+uT; %|Q9d܌G.u*H6PdiB2"@c@1==|pRR JaҵWt.* "a} q>՞iPVQ1ۇ-]oZA"7* NtYt"|Aȹci 9N>y%x><8'^.ky*w.J g=k!ק?]ZE2“r:^lI”P'02fp+]A!8ҝJpKt1xC^J}4GS#X\RPV}ct`*I%3rn8t2mW^|n8ŷŦh ޲|R ax7#?w#{kޚ7V=S͛Q3ޣ 5oQ<՝fW]:u[q8NZ;(|^SF1HjP5Eg퍱>h|\!ۚ7j >/%޶!E9ᢓ(G:TI-zTB1w{F/+ \Id8=7kn[m^ o4,'4cOJ2D4j@9Z+\FW.t4mk -wZ0][sSN̓K7bSͭ@@5yK*L!|vYâws}l^k98Wy~'U#k2[.-_);vgK%FQQshX+Prɡͅ#ʮF Ca_VKkEQL]\l]MNo~,GJ73qCFɧp6u CVBǨt.@4xSB\1{=ﲯH Rt(6o̤ɲ0e(R.k"Qbcr!t!ab2d̻,XU 2dRb8ظ VB-.^zI%1%EK#*DYlUQZU/AIbmbv{̽ {0hd⯂B˿6t+K2IV `Y?7%"QWCs,/Ab3H#F[vT,(̏iA֖,`aWrf,h(d  6` AȔURȢPz rĤ)V21R)yfz bpEL&nd/2t팜;HޜXZ˸ S[k*xvvl"P}ẉrRoG8SVB*-4YIC@LlJ!(Kŧ"QM۪2$:Ф٬(yeXIYCZ%uB2 PLY)-K2b^3v/< iZNFVz|XCv1D 1^$)CI6įrfބ?2NjN 4EjuN:C(:4 ' cRBɛK`3dIԭ =$wx)M!9MY{fW\?ҡR?9gQ^8[X>#&ڣAThJK-^zSiHꅙ ey>ܒ5~\1'lV tlTPRH-88FkA$Ohk k$eHBP T9&am&ף+rf0}h]ux^ɗ?'!S>ÏepZUr%wΟꢷo'5hE;pnpr|iX&S)~RҌ,nnFukqgޭ 8.|>e4<:^pV ή9x9%#BlI-Zo pH'ۆQۇqv0K=aQźOƣ/KF. 7LhJlګ GPdC-Pkߋ Nj˃4<=_Ï~?}XCxw=:a)|Hpzg~ w}h] -2r u\)r[ƽ&>R֠ ֿ ~980peFjbdxXAWUUZ.9w(:tIa9oJw0_L0[uSB ^H eY&-z$UtNE62v&t/InèfZٽ: dUI'U̮bM肎J%uwmK&]/Y_X/G׊$o|49֐4(bISzFILT.*!W :Tdʸu9Z}p.U$l28{ڙǍl1T\>KAe %~T@FgE)s$plQ}|KL ]`J@{3-#>KM8"@IEՊD\n*f9JV:;$PqBR\XȺxUgl&$ #XS,5$:$qP/&oeٸR!X ZHFf!W)"6%qM*{"2'>͢|k)osDM:z 6YQIT_&MŖ.JNV{L1vP;7Of!;5~\l|{#-ֻ}PBO;YOH8N6y&^Q$O.?}9Oq8;5>bq˥>z=yṭc>cɬ:W?>W..oFٻ;\ O`'/>7P/NAC~/o]ocr/:`_ ]uξh:]uhtIio^]CWV/:Z:JGztבYw~.ؤYaSGC2%\/^ۮr*ϑ}zpd=[2sy\T/⬮^4[,28{_> az5D2Ee Pڟ1Eܹe;Y|8_c ,>s;߽[)͚xKUf[5}L'{t):Le*+4mYY)ӭNu5<^~ZWE5ɱM͌1i,Ji>MUZɚ?$)TJLЄ$%E!i~N-ͫ$?ow6%uc?<HnJS-6 !;~֬%i|Hy. dMsWi:npftc"Ƥ[+ 3gUCl9(wG@ac[}"*S0hVR!;Іm>F֤ CiєRCWa<*])Wt4)rl^;٬Zںy(` b;ZOQE~ i.Lr]e|L:[ȃc'dLb]Hcv|}f/չ?oBUE%%YU31HEJ# juÔ|#ن֕j$I5,F%|a_# ZkxtHJFVh\`5 ja}ɰ9pHJ6z.P*< D̤q1M`UhC֪)WMpԤ_b&1`w"夌U.hI[5]*!oTD!nG0LeJ)h-i0mLWjI @pBVKHu1hd(7p,ӆo5ei ,q)Y'eд o ųnJʁeSC`d 8`JڬC4,IPC֗]((_ je X ` fQ\5,%+@:{H*\eJ2)G!2(N% WEB$(Mrt*S Z@e;Ϊd,̴X ߁p #6CA! 5ad\ !,P C}ҦaP6yi߰5nf2%r ykJڃ AGg#nRC3wT}̕}α8dR0 3ˁ,Q65D&9zR AE%dhdD]F")xeYyw"< eo_HV* v .+Wa1j{RQA}8ZB3TC?ʐh_N,Bh<*- )SDY] $=o aXG.#kE~B \ȸAqy\nZ1.UQ5NgWK;!pBE+`mvmU;z:ܨf\zxe&h>M OB@W /l#T Y9yH, K*#d`6uLACL`GœE `1҃Z@J$RFVWgxo8$͵`z`@H^ KZ Y1!28hG_6JrB#kUkשׁ=xf[MyY #QK#.ٽݮ ~wq2W%և)4N> ]+=A# G#A]xxȋI$ !B.1`GL y $$BLAR,Ck4JqAFdQk@ȃDHm%A펁.BjF+=ZM4-"ϣ H:AȒL\1v0IIbʪ&@2 ~ȃ\ 8 wGz E NF 9-SA4G>U?B\ٹEwI }$HA&y30#K^Qv,ue٦my\91y D /^ MdeW w^2IQ2Ww~>pk7sp"MML`–)؟2 H|,;לesԦ˔]_.]sZ75G;-4a I9QJL 2s&:J)fĥDhtG"Co]_]IlI Kͨ@C02`R_"~[q,_z$N')ڇ+d@Wx)ckG v -wR,yJ؞ #Qk+aJ2+Ch1Ua,0kw(ϥeW/,k:sinofG|h! ).JWc@n/mYMohb2:?M?ki'7;Ys*f^ڕ.C>}>Wajq\D×C}ᦇo\恊S\j6ydJ,G ItbO, >ꬬ҅/ۛVjPzvvk#Ll#0K!#uqՑR yϝFjg}hg5v'`fN)*~ [x4sZB#DZZŵ(pTNpIa1(G.>ogSEɂ.J&\]!pRA< -,5G >?8yF2 Jd!+r?U{ާdp>n\o+ow]NRh6M>>tŸ4x58ZeU,҉/%],y 0I `(!x!rm輇o}Ĺ_}J5W?9>x,WFt|0N x9+9)p:NنLbVlBcPiek=X=EϚhD$^S%Wʒ@I^J]fp9LFpB)HTgQrP۟5T*S$y}zf ,4M >m̻Ԑ\H_)Yzn&]GS xOSIkPčS;J!Wb7 ntu};/* cO5 fbr1ĺu^W]dU;UUΧia!si(CװzAɈ:(wi+jeP:|x~xg'G߼oo/Gwo f`8. ??Ln\~Vui܈].mvo.%y׽a>izҳ6ǯuǗmfl:|X*An_Oa=T^JSZ'il~0+/ܼhr\ά[cXbC/(7pٕH:f+`<B,ipvUROt/l,dAݓgG҄=W.XXrLEA1"KɐX)6&pMWM'>zlEڪk7]6`@RoފFI}[]$N=TXIUFT_ yHY;5tWX-@PBgrG ts ȷgM9cRtOY6) ))g%LeΙB 'ɗg)@(lykneK{Ĉc3_&YݳwtDa)Uw%a;?P&I0OmW0aQi5ь SۆnCOf͘8(UȭՓ15:sm rx/2Tpvg7_ήgos\2I[gά̪݇67F^|mF^;s>L'--nޭynIUkWo8t: +?SWQ"~zXxޒ\<,S]=gSY6vqHH.SUpVcwLSp6R_bH]JalL9~jg~7qxޞw~7)ۯ=\D2{*qr1^kNYE 6*TbJa+c=ދh4+M,K,195WFhr|\Ec[SBy|m[ ktvc.? d$ #K 6fi/VFPvyҷ\]2]h; ufg>KaMsx$QK{< }% fm-CAKZ[1 24Rn6j͙SL"lTCy{.>8x49NL|Q '6֘qPD%^`M/:g|Q)5e27X&&-îٰ((y6Uԡһz>Nq u'7l\T.dO޶5t{f5\iǕ~e|x*yklVse*ebj“Gu[ܷ]=RGӅMm퉏,R[l혠Jrw(^ |Zz&m!sq#wcb;NPiY!@'ڙ[<<>xVΘ@dPT;g=%Nzȍ+oL^ *XJ,$56 J7FB٣8ǬLF9owG[xY Bh~dq]2@6Z\kB,Ff31^ja@J?QrrEHIQX]yVDQ %Ocܱ;\4-^I *J0gͼB z qZ8 p_t["RzuTY !#zj1!3_y&<_{v +tuew+ٖu]vSCؿ>~DMPan1 bj#$b.ʢ*i|ՆJY}rd"WS Ad0q4 +H⷇\Aw?NXR(lK_g* YFZRIԈWwd"|4DomTD٨h:&=C١ڂGUU߲ghĜwtt]j OdN7 n]ns *-㣣٬O] `b1˓ Q҉`UU@V G|g|EHH D Dr4)uVi]n#9B=.^A~^ vxRnIu]oPm/Y@Uʔ8< \pnL)z&.`:5:EcD2J4xI^J⯆"cAOa F}h:X\LU^hbk14fu(cg3h,P2X  "w,XZ2l;z ^zDu)wL% JHzTK(F;1ym-Ǥw"L>]PT@F١[WBj}>N~4P>Ӌ: ÒΫwR^cOM^a}У_?}]| wUo8IvĽ]ġ;X,B<57coz:޸^zHo-^mɭ')kHlhVEZܱ PN.@w{e97Af$uFX Aq)lS V C^1+fb:oȞ~pQղ`mFкȜ-|%p^w{=9ϖ=z٪ZwwgT]_Aߧok1FxxCKv^:P&M'xQJԃLgfb!Pld/xʓS}cnfn~gvweh?ibzm8艰%4r{:ӕf:^Ź=kQ/C.3ZefsdP4ISKݛrd77 [C5? H=Phn J[6f/ g.ao*n{o_5~ZH=v_f>Rv CtHrBt5;\k^ͤK͖!(Тx1j-Lf_StM+ջ|,MQywjmM!Z- z .'B@yɣ&Őj5Rm3КR jW6%$AO6ZmR!L>'2 /R9!Ki;?}ͼZЀ*v=Y\yO`~:>,'rMPYU,OBB@ahl8NbAKǁS-

PASL:lșRe6"ͬrmCloaR~Ǻ9+8GDG%\A#T2 KuQHURꍀX$zii=8Ŏ7F>+U"u%u^Nrr# >Mٌn$.y:vJu Jek (խwٷB* Er W,( . NNZaժ/Y OЧ'o Q #uQ:z|Lfm\ohפaXaʫVe|aqC_:O+J h!dWΔD?.) $Y;3}rvzmy7D7W>־k0حzފt=9L|#92&8K#9[ ΔRwȜOU%E `zpsRJF/zg:0zRM%'δĴ@r.F$zZ |4dz,Ze]ZEOߡpzLp@jrcj>zVW2c+2xU5XZ);+jvvpE;aFW\5`P1J=ҳ;""jk9\Uk=vV:•)$5U5G Vk8v"+W, Į` Vs \Uk>v"+]]KrQU5؍'$scjY<)X-TZ߷ܮJVO$?>Vnʦ[#vyAL?g׏jma&3 ߏ4~c2?'4TO٧8Yؾ_S->*i~svǷB>'{V1J00m9S{~}I0 < (JFLPE(.ihv4q!P+TJω]gS\ H"C +4h*w63gZBhv OoU[P_bu}}_?vo Hz[>DZn1Oto]ސ{7bȴ>[oT P2BDF&.QBv8<4ܚd$Q1ǐ5QVdoalnuND ,F'z˻YMMo|lӓ-Y<}RNVcgG#?v2zhrg qbyJ2ikBRN^C(5%7D֕2WEV)⸱`.K[8Fp&T6YK سNMNwQ( ^ң42/;<±tG?gp $ةρF}h:X\LU^hӲYgVA;v>Y6KڢՅ ;tpH],`lT 1h`9*rL*x' .Aӥ E% 9kTI]Vw2mY9gjoFjIҀ8uX6m;SQ=#<: F=[-1GNDԼ1P>=.6Wh@rH0YA%=CWԩTW>ڶ%+ TݼE-`!*aΠ1JGϓR6)Ь`4"kZ! 1%0$΁NP\b=ɉ!d |l--̜)q1ytդ'6[kpp@_:O+ h!dWΔD?.) $Y;3}rvzmy7D7W>eo{ފt=9P|Ca 3HNq3T`2Sad}IQܜRt6EYj Tâ_|dIɶ31-Vkm%LF n,}׋V+袭PUX'tZ"T)h}mE(XP냼gV&C+rmȢe wdbwŏE 荬ecj #:IFQ y}ܘ%OƄ V!X+KR,mS &5@ײf ǹ#`4h[/63 G'4Q-Z~eRId4.[k͐k~E Iv:fR>T9Q'/z\pNd3R oLntE`etZsp*Wsy.4N>=~[TVQ$bib^$)#&7@63ϋW޵u#"m"dbޢb >me+iE#ɲO899$g!3P"4+DD)59bHETB@4e#gE9+ꏯ@JF]-bTx YZD!G#)rZSP *99gQDf0mD |ZR+=!8)1噓Q3{λ7F HKX"R!C6Y:kՙ=P(uOi|\:y[vHh|GPZfw6]7g=5&F'b$ZHVDG#~8dEP+6K3]wW#E)5Xt` L'Y9gm cM{S ( (%b0!C#0pezݦkYntS1ѫ=]͙Aa]kb!*e%`_BAFm1;fsvA4.,uxP QunqHHwD@4:&I2,*͸J1  B4Gud(`ܫ #jS\)󶍒ZbF9B#QĖsNMtjӧ#,9{X ]pc\@\DL Q7^H`)u!aq1 uRF &{'|QXϷ D=;?sjS'[0>==0l~/FtIXYnR$;·ifMMA}IOȓX?GESkƜL;`tob@ڞB#L9CWRP;5AizW oE-6`MH9#tHiu^MxhݙRT#{QVW "9lU Auu*"M?V\/ *OG=JF6#Zn:z|~x]=x; Ξ!Fa5[ݍhdF0Zay鎇 9-8%Ɵov^3d~3JmfU>: F0bZ{C=ЃIpA@JVNkuN_Y::HwC-!a^ߥb+툟S.?b=9 *kwG[}w ~xϏo.o.߼x _^~V K摠$z 4۟VoZ\MC{uTUV9{] v1B$gD|0ҠmZ Z?y™=ǠO:t`J[ b~;\UEIfXM Q 6ߧӸuR`b5mc,F{%0Ubl1f4);P=ycDf gexPQC{wRon^@:ǠrLY EAj$I~H #2yc|wxM!9peFa cx.zF/bJ/ $%)Be9:(l<+oruU3N^V`j?V f;2ı["d+.$Be f#-e8JcJ&KC!-< GE6%J& 23n23Q1*] GQRL*0!+F <UX-zgl]E佖豉hj4"ΌS%<|Ӱ#s:yYl:S,j]Ru ^u"6蹛P$ Tń>P@JOQpvK*Y~)yhBJ֙Wv>yRxyU$j*ނ귋w&}^{ݸڦ$K*&܎<Ȧ|L.^/بfs2 ].G7KInrkl=s6}NI8wQ9G~dԙ+$ڂ) -`^י+KZ,u$NVոN]]%XVW{_J1{0} ڧ?5H?j4rWȚA>,;K*Ljd<)VOhA #(H8Heg|l7.&I߾vEi7j`7|܅\g$F0'3u#slςŎbaW,%/?^We'-ˈZkI v9DfZms$U3|>-wHZnf t̖x9ט_Y{ Ӕ``/LB|Lq/qy\:o f~s%bϯU&^] J>n+L+2~u+e0b30-]]ڼ*M-։A_j54/<fAQ7IW }ݼ!xaFB A.֑aLScGW°_L왜\kg㍍?p7¦U\?c)$'׍\bvLy9rLO_}9[9pR8eIE.N~'}hãm )xȋsP)%v2wQ+^xpkzr=$ A.J@Yj3(RE9[-OG'Ua |m0&wIhUW}4avw㵾KP*m7yѣ%h*?ҹjwꉪ߶};)shZ#S~7UFԅ&wO[}L]bV SŕPG`G]%r1;u%U65+$A Cvgt}fgH0P#YQEuÓk5ms8vz Yň)?'c̱bؿēE\ak zG:H|1ve,:uqo{?1d.SaaTc$ek nww[uj+-0}?8V0#*(`FB!ew1)#-fv2)np0sH[]D‰uX6,5,[ǓY H(򅖜N)bxЄEd hEo| o,cp͸G`AѨD.SǢJTNmիQWBKD>ur|4[Z!]]%*U^RcvL &Gd &rZ]]%*5j+TWQuTC@&hU"cQWZq`RF]Fu%%Otu)c_~|snOd\fOMH=7M&8bv@ !G0y0<XYLY墊jp B\M*ky\N\?z0 (cb#L@E &/4B)n儣֔ B8z!9/SPF) 6iگm3|\8c#b:uˑx 3>ewi^zHgyI8!ww' ;c!Z'ۢUU٠e)]Wd:p u}׻{}ZX3fp/؅{L]`B<UX-zg՘IDk1hF!"-=/1ہ쾎>hϸB5w"0qh䔶CGR1#1$c"! bB <* m!,ȍ$S[)rAr\DUNKn?PPgczyEӷ xrX[V)riz2@f96?áZœK*F 'L{B NOXIO[cJF(EFJH1rKA eF/徤rkl,׌Q^ta6T̺73 Ӯdp_<吇l`&ɝ3@fj+-Kn0~K$Mے]Fj(@nBj8%XOc7iΘF C()ܒ2"!Bѓte2!$CTAc7=SSRASŖF-j^G&bV` V qHu] pYk?KV}Rp.O3O١cMj\x ]7ǫq)ckY>(8^"|D]9('#eu@8;X_~ڎdb+?|&E?%vӅJh3:!Rp\eo:Yε$dUpOgqotDT90Wճ^O'QW!aS6<1ۮ-˗,c gEfXс ul2\cDD%:y*7?E%qg|'پXTۼ3éoƒ<#Ql!:wZ3>(H dER*s]#Ǡ "Q̔6Id}2޲RYE+𤼢@X *T0NZ!mr̭׾Z3Է.Etq1KU5٥h[Q\j];1[IL;bV)VžrQnCZʹDм: 8р21(b/تtXv2Nx @ -:fI >I欬LgّZYMJvJ2,,y)?2htZ-Q̉mlӑ 8L8I& |~-`;S*ŕh!`6+QI9g'MK{^ժj>_@KܔD "I~,Γ0JFE ˜gU,]\Kgj85;[{ ]RveU&NmJƹ@V"u[xith%"2{_fEv9MMɎ7o-]êh*[2rLkJ5H\b`RDcZ%+R2!bsR҂Iz!1Nk0Uo=wdɟ0q! zhYvBG`1!H٩&"b*z Xw!^WTÉW^񐣨u%(}P._8kt*(6a|>{H  *}뾊VM %MesL`/($U>!SYIA#A hanS18_Aby)wV:.E%*Pz|1jxX2kj(vS$̖O;L +pEYSh,zbv )ԓvΔ%Uȍ&SjEheOfvD`oEǻwCCGos<2j wF64Y`n;3_>"9dFnƇR{޻{'㼛S:a-/(jt͏zr_!V"47-ŵNkMѡARj0`c=dyoQK].T:iP޺0jFoi·K}q0fϒ@9~}5zxy}[WÍߌp||#2M{ϫ:P^/j/BP0tI*Zu(]Z!wI*P$.I͢+4u+*q(@슠TgWo])Į`).ήJgWo]Y߃tB>^Cc:6h@:f7y/|bMJ/˛hu$Ql!:wZ3>(H Ɔh3wB1z vJ^$e>oK"%N@.q[ gmo=*,"2̋Eq]W*D88(B qފMb:PaALJ"?{q 8 X`qpέqcx8!"Z9_P$IMZ–tMO= :)+%=ubf ^ΎS0Gi`LIGI{iy8kʙ#q8SPdd1LgI܆ĨR9qbl)gɀV=} 0~$:hk㹺-mZ ϥS8 '?" @/%E/[E~f!FXk%Mߞ0!wq1ybQ:g RD"F&yA-?Oz^vLې&;^,hMXpH$kh(,q&^qQY'B`oA,Xlh8kWu2%" RkM,c#"H[n=cG 2R kۈ2ʬvɄC#J9/Ls扡%xFh\JB$c gwx+ɤܺ Yh,\=f{BlAgri9s5*8KLm /0p>mltYO %0}{r+ rD{tO6;n'l8nN ?_&?W|~ pѣa8"Ii7V[ ͍fhS 8.%|qK^;CmYk}5_g>W`Z,֫?yfW<̠+qWlf#p6-jz*M͍V!`@W\zq,k}XaKKb IJA۩obqF߭zt-1 `D X e O/HRzN JQRTAc4>bS qY'Z9Ӂ${ɣ# |ۄW7 @R C1sfꪹ`Suvjvޝo..qq+t2YsMZ߂']Pf 7.7dmϲ=w\z5k\z-wHZNtˋ{B^YKٞiʏfbfȅg{&5Ken_]87lO`VV=f-R:&h-r&oׄ3ڬY&mf`6#we杲PiXg"J}=}3ixz$/E oI+n^ 2TYQjnm2n|^8ǬD F9Qv3]bOҶbm3wvΨuA'r>ROqBrQ.upvۯD(*oWʑ *$@~+|@»`r@PFA"EƼ Y3R(EAZ%'"\+KKu1$%kHJ5 &<Ưٌ_QCW4!a_vĨgU Ԑ_ߟ>;9l"[s!sAL!I̓e@Sio mKկ:Mݛq]ݛ7xp*xx>^F{' Isߝȕ?׻Q~]lOۍJy5\Ԕ)‘<G_Ym3os_޼mpݛ? @">xn|6<$_O˟5TɫuCnPwujm{p`gir:5-0x÷f5iM46w ޔ8 7&kR/_~ .ȟ|a YT~={ꂯWRSgk'ΘJ"+}g|HadԜY^,oSq?yh'Gr'+R*18v1qԂ&b kD(a\nA=%6h`ɃAx'@úa1rؚrG!C2uפ, <4U<: S1cv839si>ʍp^Ѕ<+Nx܁@XE C] z*H'D&Hr4F)uV4YI*&Io ҂Ew!A"8HHky;= EV<1B(7`6E\v CU>!V:-X EaU^' JL2GCE5EҶd|+>KaZZ=>82DpF6QSR$*ɭ5D,[ U)Y*wg~ZC& 8l(?M= n'}ݻ ]y7O>;Q3wsӵLzxb8uGm A1[(uWNfU6vږ[[X=DYQ; v9ޖS5>RzE]YU'O'zg1k3FuR0!eb1u .Ucl}fr\ ɗbPhn\\,+>Tr~f/xMm}$k rYgBQ©,ù35I(/*R*ZJȅ3JrD xfRtTow(TWXYW/@s!܋>}]>E˨ۯ=g,Լdx$Pb3dL҇ 4O i؞0%ƚU3sZ" A"JX 34%DU! #WLط5Ln TQ޼plQ/7_{|;emԇC__瞓[[?u]ۮogz:*hV)ҫ(DPf7Njh#$~LLBhNDIhˈU@ E>d8~ƱSy͸@К_zR)r6h< N]u>5_Z1>ydإJZX'#Ok[n\ߩ 7[vG_JrQy]Ho Agm)RX7ϙB 9U UOCB .$V85\B%UZ3#gf\REBUX^ue^(ٜ5o;vz7Tn4| g/\c@ )() 4:ˣ“D1,0@,Xfx8 s8j+%f2{0(J>&)956Laɹ<];ua{#*I#@A>Qx"3^!#B.Wߍڹ1{eٻ$WLԽ*tw{wc=0 Z%Ko# EHA\$SeeV~ƮXF6`H\fА|.$43"0lg b6 }S-ZZ紴g0% pVE-w"EMbpRSfVnRY&5h]w!ep) .!R xxkv Q,xO2D  %I۩\:Uk^F.~lhw2 PlBk/z1гЁ]6Ko:!-Z}޻VH!He յI9Uy'J#]RmLN/#Qn TPBl/tr̡$0m֑ a܂BX/]WmdP][awE8l)֣{HvoUw.}]T§^ocmEZBZv lٙG(.0+m{"%;l|B;䮊; ŵk$wUjURWw]j==N86P^?M^\5=_޾Z]^g. jO9a@f п#p-'x,/a|uűc޷ |:u󣣻[D('hR:'yNhƯ^%THH!^$ǰG;R09P"zkN:n36 xH&)n@)jq>Em-dH>ҽNF凷X#ka0 d!t"XϦh)6yn:'0qSVD&QەDD=JI~f@{Чcfq-)BdG$<(E=3]#JTSEW^H&&AGTR 3D PEL5hm%hEMyB\{G,?u MP'FsjN89y^gO+G97ʼn=oQǧȈ^RƋq&\?K _N^dqD5qӿtt +SD >YA}ec\krMV5Y&+d嚬\۠rMV5Y&/\krMV5YVçm>iFs#5xG-{D[=b2VKVA+3heuzaU0^ ASC&yI_ne3Ӝi՝wpp_$ +ޚ n? O5Ip0 G/9=MiQq+K?p!i9䌁2J[p+A%­DFO dJ[TZp+n%­DV"J[p_B|ΚPo";ۯkCiSۯH)dchJrs(_@Gapv6OiP &dQw_qI+黿7DAx\y It&q~EoԨdaZv \JLnH*Z* yO\ycJs_ޖ[rr5ud)n&t/TsMqXpݛa}33-8suwmn >Oog@Dt IS]KiHO~&k"~i"~yH+|;_lfՆ6,M[FZhA/Pִ@yiZMR x[@]*'Yw6@xdʹ Mu(u[vpH\(IMA2Z[WIX m:g OVeZ kDknk YPI~yNA>~Y fy\{[t +|zzOF Z4Sźz٠O=[x3$ Cq"y2Dy ?C?[JcYƒC_ߨu 5[QJT)ʶA[V8"j-HeʂYCi@H3㠹C_HF2ZgC9~ۢ}p>~ q%E#tzg!w1>|fjmhz|2U*s2E-yO*bAjtذRI3~Y8ZpށNp%!MOg1c¡Q@՚8-N츜 :^UkrwL1 8f"K.TGƲG2)x|j$u$9by48J:Yl=/nsDmYD%ݖꑲ,[Ԃ݄Gy#,DpiMɔhoEqm-nG7A[cDl+6 'R@ƓDfL6DlM(XЇn "DonBחnN y\] qyWBZ,ӷr/tάzR4WY//JptQhArŜ,)ij {M֠ 0jY.AT3,4/΂fD`h˺?MP  /϶jgf' \eȉYYLBlS2Vku ZYY^-t `](&F#pRPAcKP|56MK^OoXSL03( 2ٓ6`E 悊T'hHK{'3*+#*jjqoԟ8im nXڟh`-3<=<.V0\ÅUъߴ"$ޓ rR#Ctp6u.`Ԧ2pqyRPBghT e#] ) p=5yͧ0UT8g#tY9 $dbGcZ%9<3K0UBfL<$2]$dVPM6Tl"fګ0G!ՊJ{"^HymW TKI1{g3MQn3 wm]w9ȒgڧӠU˱>:NIt17=ŪW߂nyæ[s>I0"$9s֩9&߳9W>MZ'wj7]kRatk .! JMh:ε'~3xfv&ܼb|2}<8C(~uO&#+jd6P^֖oN\rh{JRN3.S!$Ngnf_60Q8eJ>ztuYw|ÉIs;w=䢻.xVYʮ:*[Bʍ_P.}1B;`+^pQ:bN螟:޽yG}8B}/o>W?"&BOoO [o [fVrkSO|}G^p[Ĝ7m7\ ; ۬ڮ)2*MWk$JՑN f~6wG0D#_oJe,6."x X/^1:X_曘)2ȹ#i f`ФnѢYz:l"(HqPqfЍNA';Qk1'`gSJ6d Y '%@zɀhu&U ־;2[x}Cp H grcGS!GA/VhRTkព.X^P$X\k:B=&%7߁kYf5r1[5FӹH .zg\6$[Oςx/a,ch%y$aKIƸwmI Ov!d% !b #ԒcbUσ(IMư)k]Sq-Q&bt&206FO(()N %\hpjB*AaXVa=ZF؛hFs+R̦ͥx6rt1kx>JA <50LrOXhҾս޿EFew&p]&b7Ɗ>S-Fnw=V}v腛Qm\Kd.C[7&ٛN^ \L кez}0ɆNgm6Z;UвeíwnzvDk7Lv뽢A0mٍX+)g>OyBVYXz(KSgdcOR"k/%%D )XZwDðq$$ ȲwNm-.;Fݭ-h9MQ &6y᧎~*Sx?ne\Vw/+$^0N0<^B;GQ^LT|_1u+ItQ_Rs6a{"3ik\j:_i[ +niI-S'>_ŒNLSUY{b&&Q[rΜw'mf<Ԉj4<0IM`1A[n: I&MbfRWn4ƶ/杼҂NDNzkW뫙VgP+j8Nx3N2F_\wQ `%Ì@ \R#ØPeĥ=zϑ1D+ƨ ay*lybrKLpc*7v*p?q {Lk6Zug1"%L-/zr,—Evm7F<4`QG 8 2(j&Wc< A#x̌Yܣ\tҁ_bv3WqIY5h#!Nڄ>䉒MTeO6sI&C&F'b$%ZHVFG#>EPw/.ȷ](|"qBF҄F 8!p$5(W΂JsF`8֔I݃ei;P3z;m(RVR >%4HdùcF:g !jDs|Z+>`nɓ#>nFIW\{Usc1 (b9&:iק#=(xO@ )w/i$:(G nA`P.N%u=wfA%ދ}η>UuWٙo 1O0VT|.p>?N0-F8ZR)6Q0kejarc\YLV6wS y?< Ob<؇bɝL (`@ O!ܠRP;1IWw'),'Jͽ(85iatH\)`B:8x( |Ň82 G; wōf`f_5UktRNO6`UFz&[B|nJU|5y 3¯UƷӋ›El1 ,_σa|g.ц?ݦ|ͅΦ* _3:Fs$Wt5 FaY=QF YL|4k'z2sp=y8Q Z?j]v<+KaQ'$kh%$ ́}6VX)>Ũ>`KJS<uga}p `o~o{}|o?ߟ }o߿70 k餋E&{`~Ckkh**S7q\>L8rhnY+Coĥw6h6:F i; }ۯ וv|\EEߖXIw%D>}I}>}[j=j,lɘ*G1 Suv5F_jᄟwm)2K~V;G&3X?6GvkFnCEpYt jf3$#)HO\Sёv}>cBm H͐F6 Cg=F'!"2x`JbQ"&TMW_I.[Q>jS Ϸz|+D|:y:!Y`|(^hZWWu:Uj;LF\)\w{̱<8F7] eEߏ:HG-'MُJ!9a rQQ&( VJ*%xYW9ĩWirxBT}`T+![%MyCmkuٴuL`9VkC͟xW_REYsۏa;@ڻMu3'{oPuxo JnԬIX܁76#f<4HAF׫E^^GɮBT JB-(q+z MK͵i 6z„θ%"TDFƨT%<;"63LVVB*Caj36eWF佖豉hj4"f#gPM6M\f³ vk3Z)Ro0[t^̨6\>SFnw=V}vY!5\KZ ϔ+oL6WfT \)-Fj]֭3®l9?6bA9ڞZhfyØ[W7t Tc/-<}k6vx3kw_ߣ}>8GX\5YR@qsCOJHG7օ3`ԖL1]*mI_NS'{Ƽ[,aZW^qxӗEV_ Ew*Be>$爭{F,B0#BX%HԔ1тFQ4`pH΁[}8ff 6L[Y{ex%/fa-妗 Cn=@mG>3^2WHf/`>`.yֽv V/,kKۼ`xb`G]T&H6v3DfҶ9ָ$M뜏~;Un)ޯt,8\ZH5曟2|y 3:90ML$4cq{CXzܰ?i7FV̦Ijڔ *mYgo1dLj6i2z2w[ѵ]|15u""u[_Z_̵:4>W<>XQ~p›)p!6cT7+f$djƄ*=%.{!Z1F}Hx% [` ]bSlSWdT^תN_R>iRҔr^wջqq,{YKXE5>c-}wH}4{}tiz Fvg?ӥJ9ō8`1waʽ3ø%  ^И''sgnLwug]XxZv {>mύæ⯏m6~<ӠOj&}2I(5d֨󃟷UP*&f:3,l =B= 4`a2tw G'凓T*8 Lii%yV99EoC5RGZSxV8 EJY3ER(d$G L2f5L9DŽRn&x92*("1p̑k6o$6D hFfV 5T;8}k=&e*F?./(:~$eS-]-1N1sf=_`^Saq޽6yP`MBB,uhqc->Ӕ]Xg.-{t0ڷM&.{]>z^9lߊ,q.{Rvm  NתT[d]۷U_[|lm{9L/bUrڈbyONm*KH>_޼n{f:{dMa&8㞷͞n[YW_&@f[L~_ klLJ{;:wdSeg5lw[78 }O1M]U"{2}01j+ Zq6!PPEmqոD\Q5G"yPqٶD :PUL*cM!ENSI~\xsYg7{%Z'.;m̨=6Bs4sjTfoEͧ482tUEY tL˹)u[ t9"l9hW<'DB$.ǖNXJPxEjVwg?jg=MuMcg_ɧ?)]YYZ@g/˒Esa>?LC|˲D冠J6)Ƨ3Nr$*ݣj7} ۛo0 ?,ݧ=b}Rń92{(%RMUXض11߭aE'?G¡ އP:/ܷx)$)$p_Ghݰ Q'E)j*Yy.- GufH7hD}r%!*ʽ7,g"YPu5Κj_GV2ɹ\Ͼ]M/A,{Nۗ9yHJA%7S{MÒ~-HaRKQdN{E!FekmPXr*mO}kyZJMI0OTFs+ E\vQ-VMr,63vB ݈ys_C< \o <>xy9zy_Uz *ZE% ZĦBVUG"S Y)q߭d+04Q6%cMM1{H *tO ~0_3Wzo%fNIC.O͎]Q;6=ZOjEs )XV6JL4X2ETu6NJVt!'tzOc̉abFe)7l({1QISiEҲj;Mx 7m0Lk^cWDΈH#"إi7p/<%TL6)ڀc.lbju NSP,ETF+t[]͢ ޚtۻl2K5bgc:ط#9 ^ߩr.ťo+EUܼ'ξrȣyjkT~`E]^̩|w-XZDwFD"0%j:w.DMZRKԤ ~% sЃ{Fp%D.a/jj:j:\N~"tFp%& WMZ2W"ePjjoUNS+W \5iOߺjRzuU-n};4&h\m>^__Ws]xpe2ܹ±ff[ѫ_߂ׁ9aS@Bc9zەˑnl5BVȀFW@)NM;#He }t{CxժsOF}e?,ϹS%KZrm2(OG(YU$7Rg:s5U''DUsЮ&J1xZl!#PT)5(JU݉85Q;l:k@~?vuWV-]i7xS,;D#>&lqᘻ --#?WʠO)~1̇9ẇk%\լg%N&wt'Mp{rM7^.`<3(yX'#JÆkl;19^wRH 9$6UrU% 7YT<y]8p[x)$[{m'Fqk}VAW`TI{Q2JVw˳d + Q=숾뫛8U7.8,z=fɮmfvlF>H~2&*)EI&P0蔼W"L`T%VkSWd􄵔jT@gTQF* RfƮX;cXxWX`ypܮ{mWy//_//ׯUz *ZE% ZĦBVUG"S Y)쮈mVah/lJƚc<4UF`f*#uFnlF4;TP3jڣԌw B$5Q)bK(]Xъ3N @y9 Y,C݈b5U e/#7*b*QI";8pAoz5`<D6?vEDꌈ4"∈=JtB) Xfж Q JrǞF+@QCn;I{xU?jڨi'igaP?km~͟cq']q飛4㤑j|yyp΃ G^Y(tNiB  D@tc^먝K<) Sm~3:\XnNƃy6Euzyf,Iu$ 4ÙTnʻ':w0 Hy5qZg.hJcJkU a+_VG>6=˟/&eQh]XK b.IVtbmTBFRK+R d^8gwQTb$KF1ѣnlvN &ZA&/ r` m-8/T?jn>^.<%TL6)ڀc.lbju" NS9XEyWknu5*xk FsϊtoG~s$~`Ҕ ˝r!CU1d (P .|k7ߊ|'VjթrQ4TvSkXN<,Akvp6WU%VK.b"B.h<\S0{BМ;Vdk9$ l!Ɂ0CVY|b`q&{ԔSZC֗*Dȃ)c>[ehu%VƯmlpyK۶-  R0+8lUfQT[ Vk=/bюMsjf4xԳwT-,tٻ6cUH}ǻy^dwbA jD*$e[UAQ2/I-iTUM&q+:2T#s;Xʙ#q8SPddFB:,3%\4dQ'rV-嬖߾1u|Q.di+!'NsiDHtr*%9nTDNpI&pY"PR>}+*-?D#YЖK=a:`.rp H򈴔w"#ԼZVNzZ1)KM.E" `Qq H%kQXL(80ؗbn7u=o6ae2a2h%R.|7`(c EcJQ^<ʉUbr_項D'4o!U=5?{3qʀH. 6y`C5Ϩc&Xgӗ#sˁB/G/Kafc㺰 ÛTfI*-xW}\15^n1~-k;{V"J*j6"ςRRIdm=RnGO41I-lQ>p&7<7g<T>)hI ٲƜdr37;$ W9R9)p:瘆dC"N1+UMiWI TrFR]yU L+`ZKІ|>S92X)(P@6j6'aqwǚqa ͹Cld[QIsFF<\r, $꥔ܥ l-Ɉl޳jznִU8_"UI|<@Y|.jn\CAq7R 稦;qTäܺ2ڧ3s?>ؠGrxg  ў9'r s@]2cjޅ<(셙{ӥR;J`+4D|2'~]Y470J('k9MpϮ3TpfZ(N o`=ur0k%sr.0ãG0e)UE7hV%*ʾ~kuׇŏ*w l3o;a^s{I.>9R5qpуfx>_kwݯ͕wl 1G<'}YYnWvQ6rVzwF=AȊ;O; NnVߦI ^zEVba4u`^żGgWBO9<< ʻ2CgӜKd'!O}9EIJI\3Bfᨹ<Ëd?G߿뇷G~5ѻ;-:AeV`IaխOڊRuFlskSx_J++PA;  ?~vNC.qEmԇzDxAWVף8ݟ (p:DT6Uk=cB "U~]E"HOU6f`4$zҬ @Y1 ~`Ix8 $be$MȠ-JgMO~/OJ/ۨcQ)jcKO M@r+< r)e,$u;ty;ɞWVK|R(CIx(JRR<ުiXr-LVRincW'^y/4i炣ƒcĐ (9lMgobbtOZ "LKJ3NSP ?ASYDbD$b zA *%ՉKY*(ӶڟzKOϻmoz\f7LO]5kI)Ԡp:zjHRxK4T8U3zIc6PX8‘]]_|=xmr3/\x|޻gކǣS:á GMbz8*=~7t[:ݳPڧɗ(>qzls- 54ZyE&VR Qj+ay^Zh4-jF>4$:z=(1FɆyh[`QE}>ĥDhv9@E]ʞh1[[ć*e$S#։qGB҂*; &84d@Y]/X'أ<ߞ1-4FOHRV;$.RR>@0k"'u2/4*h8|b>S26 ?3wJ\Gl`Urx.S\zH/Ť~zZmG &J[vC.g{" J< p;hy&W!j@#ONFA8Du['#ѐA9ԫRj!^>Box&`7r4M^kfx$btц8ś >v p1~jEm9*h>2Impb /"\}>+ӆMڠ Ֆ!w杲`pd"J__+mix|~㹎|&j  +Gb$*LAH10ބdRn%( ept6wYȻWheRS#Fh6@n%qY#rZqW5Mƽj xMݖ{q^-K)w}gy=QRY^ŽW@ :jwU ]e/&0=}#%x#?f^v2Cry\{*mBt1Hg<#jD=2:pBFHoNYDJ`}zjo_iv+|wƂV]…`'Mu]/ON3*DԮ7I Io@>5(Zv/+{8eiV &8ɓ4jDʭU*ሯ" zK鶈> `F'+ZnlWMGO%ژ L"IV{#ώ&l2Υ :$IuҵGK%hґ@=Rb>L+Mx*`,Ffu[ Qeˌwb+i탷-{aRMoK;7,{TGHЧ$;ʋy?h& a< LlkZׯ>WMg<ͫn2R | dcࣷ"EnPI6TRRɵ D3ʃA0p24G9:>;;NXR(dG噊)H-h(FV$BB,޹h"|4X =h୍c!TG`_b#g=:GXOOtHqsTo m5HyۀftOK4l@zE=Z lɨFkolW#g7F*{xƃɨ\?2"{D&q,g2J#+"!!.'1)xÑr!`a"" cHQIH<gB dAF E4 'C9c5`5aAEo=P쨹#"N̋،<옹vz";b{P9kn^LҶB.𗒴Pء'm+Tj'mY}iOǴ" lT#~ElRK)lQ@_1& gl+tB̃kc{/8彂%eVxh6K}r&җck0!/jGeZ Kio>(T*\{)w~|[JaV˴m$zhh΄ЩԫR}Q&|F9tmIpaцYAFs}q*ő zd띯 ۮ*wDwgOVOyCɸQ<8r dDΪ$Ai[#:Ц^xR{{%yO(l#GLOg{ҹCgoW9\Ƅ8o&'bAK2k 4;8rS@7n^NFZr"^^l;ohQhœʛmi=%x4Mv6JZ,_-C9JIVкTFYF7\S##sMP5Y)Ak]JX/^H/AHVr#2Z[Id1qLH23W2C18g Zut Ԭkҵ{Fq!iGIAf\~(r6U=V^{&UG ;f5PKe֩LJ+Yr,<^Qx!%fv(5+IK]*fSzN\rneJ,ך`ɘ72B[+Suh0L)9Ƥ} 9A[@ǘ#`( k\ҿX;WA5rCHz0&P1/ÌUn К|ߍ7n-SKbW QtOXxfӉahʎ (;'Agg28E{x)r"dvJ1C/$Y{}qN1+mȓɖu1\6QYA3IU̳1Nʒq t"++X*Y5rV~b2t姐,dO `9yȀ%JHωB/5Ek%PKz2%qxfM,%G|@Ӂ4-̪ V=U+GW;*¸T"!DNdX ~ h "s ǼM*1doq߃T^ZH+9-!Z#YmU<"*%eåbM9@/`JTc{ 6R [S .VLɠ3},9Р/I@$R Z!B^Ga9 7zlN9у+9B:7,U㴑MSc/WGی_ ztO㻈4:RLjn<"وZ+M&Wo DA1y;nR$CE&1BDIE6(e_?BcvEjYb|燴3ƾqkoo'oXF"#$e4Mv!xT0]kᔎ+$4M]JiJH}L62(Q4V?nH2['UtpnH"Г U$H;A21)qا谄d @HV$UI%(ڱYA)>G\2-9B%Z0gsS-M{EdJgBѻ!v}2-ByN'ӤX TWUS^۪J]EnLZkKsWb8J g  Ȟ+gVOɗ[HX2s^:탏J"s/of6vήLZq<]7ⰟW<́Mg3LS8#?Y]gG8rzI G0xdNe"3&שEњtI2'm.I# yKp /羅,Ţ7ogǛ_5J 8(ӈ\~׃]c) ;=RZ-[ˮ}lu\\bqŻMb,W|y4KxiW=jQxY\ mj†F Mγ>zn>F $cȈ˹v{({Rzdy z}9ޖ  C=@}CwPUG$ * l! 9}(A%)jL9xީKR:b5r}|@E}$ad^wu: &[(tC; \|bw:1>Kق2*דyg%:/[IM'lEh{ץOvvDdпwgZ7Ժ}ҺY=Mxu9f!le6zj=zhɖVW{ﮞy~N[wt<-"6K=^ꍧ\WMuvzK)7)ӇvZGԗP,M˭͟6F@{ i^lߑ}:,uI:6FZdKclwjYVsOx9N-Fx2P6l9S99R!ђ 82pd+c+!EiH ܆>C2Й 61>3W3Q==cs>|lCrQܳNJ>X^yFx>_ۮe릻NYΦcԷ;OWdJ.4=1wH09 > &,(Px2=\Ds!$1<,BF)9MYer.? w')ѺFJ]3:aL{ 륾`VX2QCq(.I)m nW!>YI_E.h;{0k%Iɖ|D>JWvG^F^e2.0wwGVWCå@vpkF=UEuTV' ιzp\AG%BNZ#1zE.׼Ora?#}^zm}bs{Unn.N%>rg<ٲ H&$Ս#pp59ujO-4[4FD݌I%Y7+8F+oӖ .БjD 5˗ˡSi9=>Z=܇h|E%0  ` #I S"sYK<n&8=M 984H咱0H10&@$̘ʵT3w*W* [l+)U/}j@d٫/,9 u%4Ȧ )5)<)I7̒4mI)A(ʤEDNV9 O |dc^xG* WNfwM,zrFفWև'HW߈z-U{K=>O/ΘҀjL(;a:>td7, )߯/W1W? X@xJNJ揖M%$L4K R5B!:+2 Q`2Ȣ);P)b'3o.e =5jـ==lZK[>zqbY%[C*OKiFo]τM#;z#vNݱ!1:*Er "gT@;a}|OaMqʇ#ܾ`c ;/ӛI~}Y;+=k9>ˬEo-r@X㬟s?'Cf^ԉO4ʡwS#t![~l+凹pcY<0s}:yv柆3^̿`g77 |??nf"|48xp{y33jri-79Nj_=`P}gv.3I椁 vffAPGT' $k<8+ڃ;<, 2 ht"YP8E>R5@94Hm۵1RimBṟ0D${Z}N:9=(O}V>^[p e== xjV kܤ,آVH!)QLt~k6`x7\f@v7,z!n ~+6ŻٻjpU#\}pڼ/î( \(;UR䆯^=K;q9k];g?}γ~| ;5 ӡK!$8n6;X6XYY_qQ<>+2iYeL2S۲$'/1DBE#%@KP@АulmflXq6/'WY0ʙ~t֖!>8 ׈΄{hP ^s 0Lkn=›꺿 Fd*~dJfպ* rg~6u֝׫ Ŭ7W^ښ(mۮjo|q A\&+7?^zyf;9|͍S`bAxˊ O ǟ_)ƛyS:K0.|SZ4C UZU1J1J=k/ ķr 36WH'kA_ @.Z\]trA"Z@u> l2̜ W&&bw5O#fD}y#0PkI -/ lRtV,eANI ${Vv >^٥F5Z!Pxµk#4I6R1$Ǵr -m2f1٢zorr}FY =#|<_ȡtx%<)pcB`DŬ)d1Q^yZU$p^AO#ǻWYpG@i  AmSNVؠiSӆo6A cnŗZxEނO qgcr93CO.K=@IB;$eV0h (/NJmm}bs{Unn.N[i_ r1hMHJm#pp59ujO-4[4FD݌I%Y7+8F+oӖ .БjD 5˗ˡSi9=>Z=܇h|E%0  ` #I S"MB 1U-t/n,IX㚒!(:)lzM@Xiz%Џ:O^us}A$>E}ߜ`ݐ=V GR6gz|.)Vc?zPA*UAv t輵C1Q"aTmr~ʅJJ miڄ)2Yj" K!,HdA )B:s~S>9ikLOSyRGo'%%i"F3Vi>M H DQT&,%r4$l}r_sjN">}{%]߾Qviդ!2'1#_}#yۯ@@gR5@94HmۿBQh<82n#|!n ?i{z=7ԬQt1=z{EG:] :$D$iG3ֹi|Rd+ "K(ZŪ&%0`u^gTB(hdMX #t?BkӮ6o ?;%v>H!)QLhO1xS̐xPZox**I~:vtqF9ϑ~)90/R&NB32 dr"BH!Zr}t ]AgU_1i_1{~ԅA~|^LpMNx2@8Ͻ&:#XYmd+}1|A6*f"3jeIGKҹ7VT]#9G<Zi, Ͻ5z=S̪-+Fm-COU6åC0X.'+XBajYYʛ`aR,ЌXu;뛺NDov_+/v-vA/^eLj-R rbQ"#Kb/2BLN;WǣE{U\B(Lm#.PDJ׊I[-CFaw0~d@[b."::⯕9./҂s/ݬw]cv#j vvn@]TB`*+wmzlMmhZ :Z˒+I܋;{ Yy6bsﱬ9C-dZSia]*ʻr+-&3T\ͤZZ{?> ӵ`Jń.J+-T$%LῢRyQ ME2*t9TQJᮠZS0T[PE8W ZE_ΕfPOK9B$3wJF,%+EDih@[:mExW42}ͣ!pahec%! Ar g7<3Ͷ8ji~4K v4ř5+?UZ8ΓTg\`uy]e{b*'99u*.c+e#zIA`RvGFp&QS'BcՎu3yOZ\ȑ>e8`ϫbMmѴ`/z*62;jzxd9Jp4Պ+ (粨Hߞl+ %-͆vKTO5[b(TI& NuVJT"V9 kI25> AS^j8islrRGh\bP wJ AMEԡ{ˁd3#_)6beB=e[=/|5|&x]{|S骘yfۉʆ);$K&n %Dz*PGS#"Kfp&Z5$I )ub#FQ[97Xm[3KŀIC(2r* Q)C@zϒМcTZJX-Ykl(g\YͨUB6C˅ԏDe X4%p uʀ$D&YhS>}-?Ugb̥ʀqWI f6 Da--,OD$S=Sˏ{Ik7bS& !PTsμ,(&6:0 ]Í-1}OlU ^fŶy;-<5(ZWaMa+֣&\ʜcQ8l$A(%bDi*_…8XdKᣤ]ߛ0ʌC%RZkE5XN "x,P"/ўx̍NjCڲtЗ^sFlRMYTOm)0㒗ъ'.*#Dş2{P`)=AANjHQ2 q%U6F4XtK)$ژVdxrw.#-D T"11p8NϝV;TQpCwR.CB[#gC9SIò,׳A`IQOS Q%㜵Bxn2XɌ>e& A{W{Kb dLȔ|B h(JD(Q\_h/{< {qR:z%xn%.ha:AxJuB+g<.I0p^]}ښ6ʏ(0k ۠ 1:@Mpb)n.N%5xZ]A&.>CQUynny%R=h]u{alAgsXqg7SF%o4 ?ύ8.߾QQ??Gw.'u)< C;#`\S@OF Ž/]ƨ/cTg$r䍣H''J5puqig):ӏXy\ZLg gi.zd|qyQ%$S|2&pYh& WW|̨S>thں3iÌkmW,g Ôم@2Zq}3G۟Om/_sBFBq$.iu/ƙϿCJ?Ǵ3PrJT"O1۟t~}?|\{=;O͞^q7]Y]C'̾d_LuG?_RTIjт h ˧qKx+}BLmec$NiVǙ1z `w'`CY[uHĈ@p^mc~hG .vX%b'tYL w[ QyDI+*0Dq<:2{xMh!|e\e81%i 0 ,&+| ^µdA ;ޝ׬BrWʭWQVzeu X'J0=?L)=\V}/🩔b:uK80E]@ue`.L2ZAL%+QWlåg"7jz۫>̔?L>|?ӜYt} KDr+_ ! kL[G WFr&uuZh0Y"o]$r8/4u5i hrXd|&T ⪇ 2a"WNOF>K31#16~mợvl㝧t^"Ol"{"+*NsiaG7I6ISz=]*;Y}qأ_Zz7m{٤"Ӭ{{DmLpr;jbbr-r`{xVYuH"X"3Hf>]rzu_חOmFi-G[L<-dFڂ[[pk nm--E nm-ڂ[[pk nm-"m(ïXG!&(>`I" =Z$?JM,= ⭗h v.0fl]` vA)0f.0fl]` v.0fl]` vq v.0fl6ߊR#9i=j!`Ra%#5Ӯ.AIY-Ն4gXwx0ɽ)W΁ )ieΆ*Pa [[W֫t:o٬;ĖܺyN7w>A -Wp0X;o~,(xsa5/&d{DO-Q4GO(@5g?ki@nMf5dE?ByKG|6\hT&?E6r>C_܊oL9hSAOZ5R OK_B⚛J6/\)S:읧v4 u`s~d5:;7|`KHșbt7bQL @mJvڠPbfZjNH{e}H-6qdhP"E-K-g! ^xh}A]t1$\& M!(2LHHQԊ`|qpJn/s6}_d8#ć{NfjYEZ/[7 ^{xvm~ ߿|,+o{z+/xiKё*D 5˧-ry~s=lm?^6(YOT@$mI sDSIB8=sM5m58 紳ӄ!tnHJuC*UlD*x6{A  L;%ƓB)FAi%4 St缋NӔ<)K7ѣ$Yl1`&cIE.)n3qqqL'ܦOt~V޻ݦ v2jm 뗋+5hAZv U焄|uP߯~}ߐ:cg.(]wtH>vh7T]b<:z{ =B(1VgJcE[撌YLVIr2dB(97j̅buVdH(Fr]>kxA2U1Uk&TmNihQ/m2ӑ@%ڐxkʋ챩$XQ.~]wBɯ~OP\Gדox^&vw =K\P_yyXi s'M3֋^2+4KV lg]z֯wĹ剿_O}swZ1Oԅ/KRMDNCMmGa|oA?-[uKA g?klx6,٧2gy2ٟY_ۮK;Ƴoן'~\\ry bG!`]~n8M:E .;!՚DFT(maccYp=CA$mDIe y_,/B~f6F*ML:Dʢ9V[XHeާV_A'i&n nNɂO6_ם|<amR{KMImN}x:wGǯ7 de䅢CMeEN%c Pځ un:7'Z&fJ ETɒdBNPBf!:(X`=@Pfze݆c?_N<$bŬE/2Mo%楾u ӨDȄE-r.y ~oUglS۬(DF]]36MS"vbn :;a1=#+/4>q~aԇ=rv<=wWM~lw|52GUef2E%gn|aIے_Fn)턺Gjݽeo֥tT/D0Tv7`*Eya}uZJ*58󩊅 } N*uNE#ĔWgm $1qr}Z{UOv:mb| +WkaSu/yZSHPI9+_S0! dU"9bl(h5<(o]70){CMݾXKg o޼1c9ͺ9"-PDb]tdcɉ877ocxD6n_.򰋼sy6WXTc4M^i9m=b^hɲ(G<@jН+{ހuBM]u. Сeظqbz^׊V 6)GSKx ˼ b1d:]4PtFUKV{4ks҄k*}-1_&1SwaOXBTσs%e`V&)CkIuI ?aRŖ-ӏ1]0/Nr3|#׸J50Iԩ@'%ۥuA'EűD$Z!z))2Q)tBޣr1EU/#a>$m+P}62b0(FWS3(3CumV`Pk;,vDȲ2e7$IUDY},Lku Ce}X :;8Nfn5tǩEs3 &Xl]2I8eԬny go}PRAumOsr֟Rxck;zNS{CՏ_Kئ{gǕ"+I d$‡LD Hr\si6\E碥bT\XX,,DGJ0HFfF,Ex*Xh,<(JTmn_wߺ|94-YAG/<2bf*Xc hdlE!Hi> IW Vm< Y KBv!ǀQ?-HηNJ݈|ϸ<n6;ڮ1jmƛDŠI "JHFS,N$4IZZ5 @Xz#-dfIc$U|L"b1b#fި_fŨx*l~<}cD"mƻ,ƿ JeRNHPgG`EW5>X`""+d@D쀙 fd ,Y4k[bKZI4FfZQ#[()u6%OEAƸ\pqیs BHdc1A@ڐ2Ǫ: ɔfSP}MMg_#wc]$nwrTAK`wHk:#vEw@Řa{@S րA#) [y)1z 8-uzhbAB6`΄/,Σ\B]sqigeevG׋Ux۱;!6NW4ub%Bۻ^/,KrUuǭyp;:m݌];\ƻ;_d٣{-χ 1wϜz.:"k%EݻjX욚_66<7n7$acW_ 1{m_67BI6ے3r')o26uImB2e{&"RI-u"SNGC?ňcA J)|5“f<6dt”[S@mp!I6+v2ӐM~Q:-'֌ZKK1My>~-mgퟮ.AO.Ӵۜ춮k9xnwc($"yr#[+M Z'' Qu./ 00*T mOx!E `!+K)hy΃5$eCE&%ǘq@yy+>қ)3tNynCzb~fvS#O /zmF)rF-v*/Іَ!4,i<דEg# l!+Ha+Hňш,&Rl 21%̀{J!|VAQ0PtZQDm3e+lT)iBAi!ؠġ]B=.qKlėZxUE 7)#bILlAgzɑ_1yd004{,0Ƞ˰ۮ7,edʖ]eeJɌ~lA8 -$,l9xE:^ncvNчw8\<6,~?[˳Tz``,bH%$qӥ<ا 8"-bhH"W_&ӹ,H_vV^~wN_0x 4!z_/#J٦'|۷EbV*Ih BHɶ^D,\CQ'Nl_ӿ}N{6(~sm e a-x#b[\S*%"+%4z:LT/J_\xgp!J-=c>3s95sOϦד/og&_O:r0h;ww*HCpN~(6DuTڅG+k;HJ՞Om%#gDWSRH10lmU >7xZLOrUpRѡ56I4k*uʁ%֊mJ.Z8g_>&joP/b>0/mv2jSϟޟ@fZ@̈W |ˇ W`ue:€@@-Q=(vW(?\Qwv,􄑒:()*KGƕGݚg8E+,ҷh+!hR QEV: cq*V=J}aaz%PC:"޶2Oe'!֫mc2Euѣ)|vqПnM#:z.t䝍+tFWAHjRMEYf~o~wySe(oW/8 ˕(̈́%WqҚ`㤄oc*Vb&S vfgnMNf`ͻ9z7f{6eJ#N3(ũxl09 /ErlNd*v_N~5ϳmÓ[uV~yZ=W\L.sN.p_~2;-wtRȹw&[e':m긪dBU*1zRm`66eW:Dʚt !j3[{Ǔtg;녎[zuGOpҖv,}Ipdj԰kawS.` q W2H>,NdL5!*R*v#܌}nM*J"bLndY A2XyTVi4dH# Ck5t3th O~_x'$>FLmua][`Q)7R$k]D3yM=n5`U?^PQ7ܯ%liq\6ztͬ,yqwNXsT ǂ6:fLjřLT9D2ZnUvI CQёvb5f5k˫dI;W! %vl]YP;E!_| a5Dֲ/jӥW^$X)PI?&M9b64*A|Tڒ@u6|-`LTF\^K,sql/UqR=c7q{ƇQʳn/t}}k꭪]ߴ wl-<2͋~..>]=}{lU• Y](∹xhQYPfUu TϬwg$@ER1 ^aF\F'`f*{g8=6]][Xv7k^ۏ^{DJ1RA(*E`WE \.VU 9Ubl֎G:jO( YnD1{5UvMN#7*TO?&v8#zk1M ^G =b=x_4ʈTE/*rpbe[lBr%A+< 7f ţ 3TH$;vڃX[&mCZd˕gGQ鲒1uvӒzm.~1~qG]PJD!D$> z@`[',8(50-ȅP)ʶ~~Xa7p}¶G|͑[W_6Q>C$f;Dōi-U*gR,OWţ{fy֡CČEXՙh̀!`h0eP~9򁨱]FCėe{뗻 p#/!`6Mm /O6-~_x~:9Y g6lzgߗW.7@&*BPU\OԔOBi+jKitԃS Ry R=;||:@)gAꀜ` 0ipZ/ZJ8IA'q}YCuP ; @ L ڨ:Ʊ3P@'WǕw Ύ;]]V[ͭ־SSMWty1LdD_*S *{sC.+Y Oy?3^i Ů1{oD@?ÒS>hrFǪStCLɦZA#y*S/wٸ}*Un\T|{ֳ!-5.h".(R4i;MJ0cqU=1^].2{[P!XBmN˹;@melYC+ :`Vq KqƶpTCH_Ibx~^5L8%Y\PNh bbbӵjJcPH P?k+`[m>s6_ʵ>`1aŰf;#3}fj5* ^,u+U՝Kz9\^й񈽆\qZCDP?ol06q8&!5'EKdК2SdEhm5-R]CbBq!?h+rp#qUL_E,rbnXwB)&WQ,cZ8B?f7;^7Bs9"D2]ħ^'QhxG0[{2Uث z&lso}5kc.M%.uo.h6UnȜhv!wbRZ!ESK`2/:,f- Br QEkFrY f^vL>`u1TB+Rgi]nHMgrnW;K&J"Zp&&QTP}pTκ+0^>r)lilҠ2TЭmH|1B@x+sNņ$W!(P5&>_83?Vrݚ#X]<= A0Xʶ46HU[jZߌVjEZ#UMM̟BH&`Xk2QDBѣ8|}X*wӉ*;{V*^}ߣ vS]E6E2aRڬU҆Ȱ%тU{qEDdЊyȃ>TdOӪ-?v1hK4DUBWmRΥ ŖM%18G?XjMC:z3mkN)a=$e]1$e$Z2 a 6y` &`Ɠv @&Rͮ=`E;DqZ~:PwߝyW!q:F}beLv:*M!?O5cZ}5H0Ēt̻p0p2Œ{re]}J8ckޏ_[秡{7OZ)\.\ %V–}54^<_.~HŴ8aDX,* Sߗ8Wk*?nh%ˌvwj[}YkYn[va;I9 M{W%%ϸ$!.|Y1ly1F/-faa0l1 [0bɇYC'+*hi.uVe.aץZ\pr֗cD+ݱ/P,gL2ǽп8~zo;=#DpBZf6`aEϛmEY܈M6%k=mRBX*/)CBOo~ϬC qw\ͶAݟt=gWϤ+ +h6Nʹ?iU?ܶJSs<bA<,[~w6E̢Gz]b/uj%6&T$= =$SphcVN)td!IZ^qa 7 SQ 9RTHa^5${LSdcgb$!;{t9r^Uc' .]7}$$[-v%!J.%!%!jlU6W}l&WJc*\}pe6'#+7pR7*[+v]e+p •%B}bW`?5 \ekγʮz~@R\<;\ggW3W煫Ykս ˗;Wp+\=Ki AݿË_r9^wvyoqMo7sf}?/p[(aFh"I魗L|ukq?87>M>Iڲs VkV3w\GdHk(D1kd1m%Kx:UW_uU翎_u:UW_u:U翎_u:UEgd=#i:UW_ucU_u:UW_u:UW_u:UarH+-hes;gҲV[YҾA4J3xM4qΔ.d)V0f|#|Jv.8j,Yd_دߜ9wq+6X0lKaF혎[l+XE,UNI!=32):g*OS~u LQV`J&fZ>K~e"1Z(OJPCa9.tؼ32>zEdtծ.Yli7Zyx8s| SW k\<뺬Y/&mKM\ixNћ Y\O!e n0u[Rs0^󩳲A )Wp"՛ uޤ#lqw_]qtvAoߺ-8.GGuq.q}yלUt44{O@cSb?0To$4 =4+2M;8K=LK\X*Ï|mG\+ ըh$06z M4VOyg:ryY_1FNx[`QE}/Dh>8@E]n3K6¨~e CqUpI$F.)CЉiGQU'h Lpǀܻ¢0 GXg)T|!m0b|D89$qAIYp:QA {a"Ot݅7*ό{?꣑:%cg2R$06+^ a 9⸊QeKOبA`I-MG :EJ{v}YFK%@<i/ξrQ;5BO d>9!QoewXNF3HK=) l.o^]ߍK;M+fd&L'@5!7SrMgb8jwi-4.i"7LgiqؿqMי gYܻ% t$QU~P~9mTLJ'ȷz{+h&xqLrD G# MH&;P'wl_S} lP|"KmdD׈n aSYqMQ5B(Q('X"{`HN-[(xG- w-G)|O=L.~Sq]&FNGrѐulU#wQ !e`{"QmNCv(ݭ\lz-6LmD 8Gx#vSڄ<)Rcx G* ԀUzetGᄌ>̤9 K|7IX-ԅǫ}`G%&?*;Pդ&iI S3UE/̴oH[r)'I"5Ad+!9eb7\qt!==f[8X)mM%*TLAjAD1TJf=HkSEr!UK4ᣱ&A omTD3$_3gv vqQ&3'>x3Ud T .A_t +;z&v|1gܸ3l5l O P&CG|g Iӫ@=D)z2JdlWMGO%ژ L"FVZ_vKInDJ")!d`$PO1&$+i O7hF]DahiVV|[`A:cAwO<1JrV۽̷{N?EN@A9maAJ Ju8EXH,y!A$ieӒPF(Q*, Dnn p}$|mŢظ4=5n@>Av]ǛR硘Ȗ%sc==n6n'Ğ#(a*/i"jbFz%oubg;9w(rn7X'(VkV3w:ʤ%sDZC!$u` 4*0ZɨyueK {gh|`?Fў8Ņ#/ iNQ+!b0KQHkb*rG 9e&A Vyp$@&w,%D3.1yĄqY?Px.#-E|ν]8UW&2/.JQT^Hޛhp8MDب-^)b2HLDc^C™2:Ř'F:LL"r `tX4DEo\B*YJ1,,30ʊ[WM.33^r${q]Ɲ ? #6Ϋ'4Et<:Jx\3%@xW٭ÓfΞpwDtP+5CH!`$ď|gw HЦVG)s#₋+]w<ua+]tQ@xA"Pa7@$4J$.icD2PGPJ=dd4;F@bLPH!jtA;W[~HqgTZJ")"bEU=^G S!%")bh:y{8Y>;e h03eS  Q d3>z Ah I3mA>_ʜU{Lku:M"h+.V\\y1#Ak2P΢XbqJ ȃ\,GxA2U\. CZak;Qv\{؅Qю~VyܘG]~3GW11UowC(<\j}GYKݸ6jqV0qZGr)=:qҁA_{/|i ~CA!oЃk!5BIo.OMF7˟_L[YFghw]wG_$erhg^iar4w ~crrzpBz{Wԋ#7yp)&x3B?ZQi3hw=$IXiW|̖jfno9tjK$66caN/1 fpUQYO>/(M.5b]c](:%YhEk铊#lm|%]SIg7х:|kK8J*fjڰƃl*v TM OK\ſ~8LB0Gp\i'='ģ_ 8G{柣/]L7q}.WOa~3 z*aesTjͅ *M7*K;6<@Fn;?~K@0k% ;ˌz)|g8@?FJқ[zS{>h5m6I(9pagIJ\FεF@U&NUX`璛{{-j'T|]_06&JD*x 'vI:U9k" ٴ-^MxrKnt_?k0xp {ӹ6 3Np5 eI)C3)wd}NQ(͹yw*wٸsU䶊*jop89Eɶ1WbZNKiNC\EohJYQiם/UJ%NE溠%XU@Xכb]*:t\#yIYR5c!^c(K*!t\,UqsdNkZjY[FmnӔ%GenxwлŰS-fF^IxJXfLu| y8^z6i=a}c ,MMAVL)@,۲918c] )+)%-tV%!' N*a2OXDn3ʛٲz$]Lƿ]S:6}|19u!ķ?B)w1urv:?gm%wRt^."xܫYKZT:bQذh͚ND!ֿlΠ8GE/m6odT9ڥwɳހOYPWkr_><1y_Q V<:Q7פ#9x`<Š\RzQ}s^d=^b:Y҂/4Gt*񤗬j/K%Jl zF 7ESť)$K5&dw5$?$ qGC*ኔ-GHf-\NzqWLzq\G8$*[s0pUz0s(pUbH\wWVZuPpEk&B \ioo)5zp\SUݫzpU}PÕSWoԖ^1|b%ԯ WׅGI{z<Wίސ/x WIyO]6\ɴ%PqOPЇ%}3u]?e >v~LgeCkbSVκ:R6YBM!N z4+_c "~^ =e~önkQd( < f"TI1ҫ-L ,}i'OԿsj/V0M7z Upv݇1dVw~3_#SkF>:C_q%um]j'ZDskԊ8KR+Z,KiUiݞRq.IKJKRkZ,KRkZ,Ko ݭRsnZb"KRkԘV5QXjXj-ZbXj-ZH4PrHj%CI)*!$\ВCOrp#\:µp#\:µp#\:µpm!lgrϿu>?e^N=,BfF9씱St3]MBPGT|ʾӞ8;RߡDDG&B ΠgY)d )"dЂ Q@Wȡ鶊ːqRry>W` SVgK^n2=7sS<~P 'oȐۘV8 f`0/P $SU #`n3ݛ~J %"@FBji& 8* nB']8IYxD^DOnDbPd%DS_&B.hfPRs'Na- %RY O!%M!-)P.!ìUE)J Őt,eBN)J觱6eɉa)6ZbPLD* 7sR8#c=R iƮXh*cnXX<0 IrZ&_vz5]~Bt%L4ω\@h$أ*̽"@DD "2+$ AHQ`SFLF&@LY[0b(մcWԆʨ Ow.DgE N"1D`DE*ix,k}.U :r'P| VsE21C^(kh  G,$!e$E0NE}Aj+"ʈh"6DO!q,  JH &19(IH [37F$wDLLy΄ȃD$kI 'CT8[J:EzՁpq-k:iN?2V]ņi1(er}I $5T(/6!JJPs}jڱ+xޣ lotWs䃫 ~a}< :яi)TGT 6u,Ч8z_eJ;+(Y]pO] 9?oG%oGY6ǢVtjCNN$8q48U7d&X Q2XRuSRc2tX]@Kn'##VAybh5qt" ݛCB'䵭rI {Eгќ0u?lJY~-i?lCBg*7lqIqumSCLGEf{HЄaL5(6Ucˋ}6;mʏ6\]Ia'v! >צr!!d,i`̊+"$ʌ^(USƷlcN''pЙ~%.kόe ^qNIxH1y3۫lju"L7)]JHvHH8e@B 6y`!vՃMqX%f '7 9ESYP]2s_-|^~<]}O) x@0ٔce8xD"]hCRs'@4TChW&I.KtQG*xmlFCcvy5q, }ZjfvC|rܸQ F˨]dU6B"pB^<.fpDLQ]^NK1[,JҤ(2GV@pT^e딎9dR:2A kB-Kp_VC.1}Jjic̃Z!HerN4 ?OFiKMM[#{VKG!y-Z2(|@x J@zcY3gȴ\ M)ʈ=,hY:DDt>]B).g}Nh5C}p'7It6bm,lͺ0hc`2Gë@Oyv~6Ȉ6U|Mw]uZ^qi!ƚ?ƓX.}9%޵6r+ٿ$`1;3&Ld^|ڊmɑ[nɒV˖0[M,YҎxlXވbRWT*`L^V3׿L%.$Besf#ͭe!8JcJwyD]]M4DFp-Q&bt&206Fk({SJ RFl;/KH2zڇG}XE)c̮ZW^;!W6P$"TVdȡtyYzuНJe_]ٛEP2ͫ]սj+`!5yr7yzM|LJsӳ|MU2K/zFthN?)tCtyP&/qhO:㶰:lsNSۼBԙ%O$ _|@IV5oCțuGF_7X &/)e(l$D o&-eەvrye3)9O W0/  YzKf/Yav̉3x=8*+>(.!` gfhUtZ/[n.;y-O`+Bve?~SlgeGՓ_L5yA;Wף#xA|\Y:e8a1:ME oɩߏWJַ7у? P,)шJ\ 3tZƤT^Im,ҍEsd s<HJ{t ^xNm/GSRjAyj'µ{N rJ# be}0z)#"b1h#2&"Fjk4HYrXH|-]QeQ̏ a<^^Mʭd<'^C؛\ޗݗ^ 2+ 4X00iTϔ`M)B NhV! "lX޵0&``g\L@ >~O.۰ (Gb=嘤\hGY% htP΀ִptu;詅y^-$58*q$#c\(4Pօp:brZ"CV'uӸ?= hp9~{eJ;&As)c:jGA @d)^RnAݬom`d5s)h8$2N-]/#-6@,o6Bn↷JgυZff'!.O|7*]bY^s(bcuUbѬEixw_f5uS$? ExuZO+xհ S=ZҊo,Y@'ϼd+㇄ѸM"&@5`$B#qM07;C^t^h0"cSsҿ:nxq .Nzn{Z{7$ϴ%|)8mcucB. _fj/C;Z-}dInf+Qm`W=S4ZK`?ZW%wԒnjl-*׻_nb]M)oU|~=/*j_3se*{l\ؙ/-t2.G-7B[A7UA59 *q9Lۉ<7_TÌh3Qrcw4 MpY2 CRa3ˣc'SVbJpNٖ|yn4-%p gc/VH! ꘼ ScGW0ҁ^+F'Nkml<*cm~]^X  u.3E;P2hkЍ39?Z9]Ry.rɕ11"a4hoݺ9 GMZcX{(@lXa |NI$Q^b\H[:OX)Ql=N;[B7zf֡ĤМ-Si^.<&.ra= 1g}r1VmA?5r{zF\mVΑ>_H+D[LRc?ȹEF rOb ,Jc1F1sɭ`)$ŤcKXlS*HnFG-Bô5Hs.gHIetLt,RUX"|t\2G`r2k:z*U.dLb+t@HmC~sVk/(Jx->UlT5qP@!FYVO">qЁD5ai>5ۋA]-;z&vDOWissF9;(0E [cI"!tG6D(xjJ$) Cpc9GRQ Q!cZZ=MuDjmWsL(QmGs8Ȣ`;X+\$6%D hl5qSE%J$Gjᕿ;nZR0|3Q&5ƭS?)m^|-P6P3Lqng9q#XtTK֥UѴj+O,Df q&cQ{`|R(d$G L2sf; %$Wr_[)6Sꁧ[κ'a}w8O>:LK$eVMf$.G(I+gJRjezhXS"317pl_*IH WW\PLJr/pd*IDW\1 Fî9\^!\Iw \dZ'iu))F\BR' @`L⽁$pa\ҒK@`ELҽ1\:\%)Er7T=?DRJt7·'E?]8 KxX|\\OEq.OL`p &L1+mQneLJĽ 0z^wIcώX!Ÿ&.}f*0iRaäj?y͏V";/O_2>jp]`X'C_04hrL͆8M-'͕'DG#B M;+9s E4F%-3fCt!V3G=B$.IZv񓔻! >A  EdcÁeNfϴB0S2Vc:c^_ϋR ~z;3Q ^!ht /:13лpakA8w|,>g; cw󤢉K6f9\;}YGrϥT.K>%4dwn^@5}6dCWܞAvx C;0 \|wq-S}g&yK8[+R#y[3fY5<|0l =o9M +fnf;\v(Z{6,%$ *w=;I9s?*8Nՙe. *HrZUn~W 5D*0Qt6w K*R>j&]x`MhgۿBgr5EQ:Yΐt(zf7ӯ̕;tJBi+RTUO&:z[3H0},o;sūU;a{)q7yZLrg gV-5 ́kUm7s5l{{?[P@iwwD˾XϞ]Iρ(b:&LI)8JH63M"Y&FK,Ȝ:2v2]{<*kX1l\kؾݶވq n<Ytk:y˭ϓӳax> 1^@2f=%NFAfW]]k]kך[-f:ʒ/XhŔ: roC.x X d!(fmH)'0 g'_Dn3ʷ6Tg Ec9j Ԝfw4v}>`zfrr`;siޕܓqRd,iR8QGexHʛbZ6kz[U-8:s;s$ CQ،c'ʑ\M4zv\VsKfg־+`h|"di`N~}ҏtqa5RQʔt1f)S :@%JR=$-e(38ё GЁMV Ig3,ey2HK'G4_-#@RːqRry}XQz Lp8ʕ.{obӹa/t [7[*Y^RᄆL!p,Zf m}&Jdb73"l%%%"@FBji& 8 GMj[ Kq-q}=˖mꮍJlO7;-u&7rWR\[t>,.rlMIel.yz+jVk6Pѓ[RzQ81::Z%XBt R dnOc!C`KeMYrF-EPLw"aL٤TmXM^RVӌm} u w ]\ȸf9Ŏqk'/ٯ' ,` #JA2@%@G-Ud{kE?@ &g2+|L (nSFr9L39bfl̵Y0-@}մc[ 6t^`Wi΅,A$2CF`\Y%Oe- QŠ#7xN?em`5W)2(F@&!FPJ2R։LZ_{O8agǸ/~lmeh;yUcB>R1B ad0 S2Ñr7}9VdxPQp@QA~&@ - N2*{jlPԸilcbWi1(e>X$! I:-MZ u.jڱ?5Ἴ{yex̑kW6sA64 q= c?>1teGA[av.gd(,>L܅w߿?|ä[ig&c ^]4&IRKPM#Eq|Я0^M!a@߸DzB$;L:y-_ӳCryMi]毿nޤG;`6P.\= zovzr{=zw~hz!|Z><6.GsQ4H0 5^ ж:YJRc{L[貅[YqleoH}a+o)K;uTnS-Lx$8AJNU7K<3ʺ\^kU\ǘ/`,J##VAyM5q6e4y]h~d:R>)Q筒C(Eg6'L-2) d-dg*8Ef{Hߴ⟎*̸׫.'Ͷd& Lx[&dC,[r^pKp[2Lc0Rrx`|AnO\l^n<#c<,"uZ:d?6 ?=Esa;?eak{zq%^z۴^.~F4qY&0ZMF4W|*W>^R(лѢ۽e~m~ dpiҳ]B)ȇo>kOpb&1Y2^Zk:mFWcEUyOfpx(ҋe&?\:z|]tpqdD Y#jmi*DZ&YHXlKїc^W&7;bIB ϓ^"uOo޽-߿ozC|7߿{M.M˃6&LJ  wWTYҪM5[=rD&{|\|fm H.{=L Ʊ ]q -LBŃtw=L4trC`܋"v[*OԸKva@ۺBӺϥe+1Xb N&RJj)t$1ܥu@D3 !9OKQQ20!5c8|ɞӞc!/LRLP#D9LY#`) bVt;U6vpX5x!öyD[.%qUGnr\eJm)veeAF+(9MY x3yDڱa%h#v3}Y$hn,Ue;3(uZ|9Q:Km8 (!j0 @ZDy%,wVzw|:k:fMmT}xGYFpfhAwsJpcy!tzn )xw88qªWMI]~;t%$7<~ Ͽ?o)uNHAO7ܠP?B Y4I$Wl.Snse.D+.pk]^"tk:RN(. .r"ib|޻OI۝!ŦpCz̠~=@h TRbK/- f>;"S?EF$+Hv}RZ=(䮼$VKQc`prj ^ 9}+m"%$UZSo*] 7Eb7CX'G!?7F<Ͱs`i3Md:ɤDJF@\%Չ^۠^&~QE4djnck Ug_hѯ^% y@XAZ0*[I={dN.Dl_d ۃZeQkb D`QE}NDhQ%H|BOyplU<ҵ~ϯ #=^Nq!^[U$|6vDp.B (P{9]BaI&roIN"[,Nȯ-p^Mkw7Jv#c0 ^Z.3 +d W)%ធ Nhv$*-%X0r:np49;ѳT$ϗ iK5M%$TƁy' ('*h0!zC0L$O''.3ݟ·D]3KH![V$Bds8"DEcsFv\oa-bYoKxv] !C! kϮGMZXɎ.xffmɤxiPAA&퓓Q:ND`yD0e}$Aډxh(FOfOyHКVgnێ?ʓXG_EAաYͰ-3 N֏j=MҰ ؜ȵ`8{F{Ump qkGrh2ZgX_"E1գ92<a4>H.u?dh]]l[,,.ܼ\ ߼M#1JM}dJ_'7L5[nߌD8sfseMo! 8Wtɩ2gzp MH&YtCmv7- _^MGwI܇-d DFhȱ6BnF%qY#r^'B/S`HVZ6jf^歖6P׭]@ZQxN+y "Ijy+{G2"<%R`eݗvt}P[p<:Ȇ!gDJA+RR'r D-vȖgCۭC#q>; ghUb9 h P643QSb;5mB/Rcx G* ԀUzetGᄌ>l:eQvBT9iT:]uٮi]m4`*Cm^Vո3T)Ua&~C CRT=;+msc鑉&]>(c$z<ɽOrǓ^ T}ÒP+:LB$ho٨!̛䓎iy $Ȕ-~˫9?l^QONtXD4ŝp[kBGI5g19IZ&fDztԣ*RelףC,-5ztBaTeѳ4!^rӶ)pamr:M2{z>Fٞ%e|]~ X}n-wb@WɦQ5Iyy,Ҳ’R YVka~|L9ӡ]A(+JO/7̰/y uJ+8_\s_\suwbnz{o=o~7ⰸtДjwdl.lKE*(& R$y:F>HJM1e!u[h2Z A!Dr{!{KE@XDKSS碄 ö ݚ-&%O//IψBX_ŭ,%7bE~mm/:ɡ* Gh NG8 lj%ē7Rэ@qngu:6$<|4Y/&ҏM:HW`ud,ȽGzJ`Fwg$5LYB=S:#,3%\k bumc> 5|l&+ \|xi8hFDGm-whZŵD(pTD*I&E;>~)5g6(YЖK=a:\4@:gGKKy'12ٙZݙZa&g;?o]˂ise) ` 5MX<* #y计ęѼ^U5,:GհvYb o#~ x-OGU~,PTzmUj{{m%WRkM,ݝevx"]EE:#G25RuheV;d%R.beT AzZ9,H]gњ;SnlSg y*qJJThc:Ie#&x4C mΕ\ 5e#'r3s *_X ݣ/ES$Yη騙 rB4I m HU&wDGc4I!84$qYœ/r7ϡI:~QUN^fʺ8p4+U L+`ZKІ|Ѧr"+dF3P ʏAN0h 'lcmpi#y*JY(IK)KOg&#*!rGsŎڟK5G<:ᩋY*J>:" N;&RO h!;FꖑsL4<@Y(O ?P>0BEThÒk9&]UJC?59U˿Ouj_Vs??{]'0 Zt]e=NqNMBS@>" ۿ$ZqNyw1\xbbl?"V 3K6p7Vr:E7^IUspzFtwIRKWfpgߴx,:8}wӭ dL3.IȻgX6 [>K'#da%rǪKAlEobq.s笌]YU٬\Zn"|3UWcաݧŰ.b{Keu 5ޠ~\W|to?ޞ/ߞptߝ}WP'oϛ!`7@qvjn:S.WlRBXK^2-|+b Ώ>qȽkjb=ՙs n-\g:Eh@TX}f&'EDA qf@gp1VH7muO(gzˍ_36a7],dL Y=%O=ţ#[Vfe4`Q9U+#{DI$cO=75 hĐ<$CYQb5l1k81Sot1W3&dcN$SΈd1FEXaLщ`QW7ΫWZZt\ZaW1Lc̎1LB+]=i^pOkwhJWthJTkբovjo{<[ T5P1XJr;Sp_P$P%]TĞtV%w]pdTJ%L L"w\c#xsp1}@=+K\^yum/ѭ̓z=c8Z~[MֱRX^Y0ZA_hmeMᗦodO^l?CW—yw[> 7=DFxYGMذre7[a^,f)yDYtӗEK!8W~5Z[8JC, نw~CG-9wd%c9-N r!+9 Vr/ XFg=b%Qm2'2/`JʧզiGs$'# muę /J0+  nƣa-}M2٭6혅w얡]xG7v̵ y]'qjQx mz&hs+/shꆮn_nfWonpeлkx7ʆl:i}M/>xGVvy067wswyw _g2p׆Iw@Ҫ7݉|Kw [M5NL?gxEc):;t(NEֻ_ȕ]HGH+6ׄ>BhK:RS^\n_:n5 .I\ŷ(S=}f|m|{>}9W!t7b24)Ljs#u;&sR!tȊ%ׂa22W6qMg(ys:zz-ӧ3U~n痢/UhևLk>0bz{UϹ˅ǻDfݟ?Rn=֖)NmN~{YEPACzd;hhZ < (T묕ʴ'R2m i4ҹ\pj%֛b|2hЂf29,N(&+6Y : 9Cv>D[üt$6%ۼ{Qd2Ȣ@uɁr2d2c9ꬵNn+gJv U(׵Kt@ժ4X3%}{%^˅z~Pb=n;L6c@Vxm2eЅ\kRz2%CLZh\!ǫ䕋K=>:.\w1,,5,uYz|}p‹2?.Th}60Jn]hN9Vr;`V8`v oOsEOxP$;(yP( >,aR*5Qc^GRգ}˜$G*A"x sIbv&y &T>گ6_c>tj}_نRs"C]:W#k7bn={!23nKV6i5s>pz8}OMO7O¹cv~6_3{o;?첏i|ӓ8x\ KB“',Iry/\C0 Y܉SK5Oc>@vhK۫? %_ UjE Z;`pѿ_} *%|:&H }q:hƮ'|]#Nz3 -[-mdsLv49Eras:ufVr_OVi=deLd|ןsu|<_w> &ph9^<{f3 e\tQy ,Yzn(Ԛ )uL,k 'FU6J(Xs.f`駄M2kgVvB~|}6.K?jC׊ GWo=/!^nDp2K$z/_RSǣ6$FqɛjDԚcx^[k*%!sϹԷCrI&VQ1*KKkSd(*D5ZAoV }LO(Q '6 <>JXߧP>4>zLL}Xܝp8m|dg~|&T>qRsWp)*p@KAV 8@2&V9mYP='9A 8sAeȺՊ嫳@t, 7G 3^##(7.ur)5 [u^ԄE.%S3=gdy AȲu$Jɜ8RGex)$dkLBV x(Upu %LCvQ[[@J# 5tkJ; x45W_=ߔS4 }ou<2Մ4#4m14swV#E@fFyl)43fMf=|D\]yDU?PfEDO* NkjtZ k,"O5:ٺE cDhGڶi +\d9dea*զ#Gc)b;|(HJyAc|0ؐ)׌NxiyuJ,LFUç! >Nknz0JDT@Jb9 2k՚vE5ldM ⸸yZ7jrڻQ ۶ej(->ͱ$6IkrȭڑNh{%h**o9-r1;zQ*Wz` s>X,JPzfSTBF'c)etM,9YX c)nS&$h*XLL UMg;2ֳUaa5 Me,=c]big4{uvr;ݩ? WWW_&al,Qs"DD#Jk>j"<8'2}FdL@iMQ7ɬJ Bh 02*or̢u#kMg;bh(池v5meԶ=j{E/22E)x,kCtUe:Di!e Q-aMB@! dRT τMg;~Mt%"G7IMKJgR(Yo c HO.Js WV{2LLVA8 A)b$i[$&-r3DUYlGįH'M\&%{}Up&W2FkU*5Ol7?`dzlh],K.r`}STJΫWKځJV.A2qYL9H`@уm-ʚƜ0Jfr]SCd<'a3Y:o9?cW~|ӏF+OWw'r% ,[0N?6Y1s!cVnT<L-Nrwrn]R;kؽzZU&>l=aOVu|5kJJo޷7WIQts{s|:~}hziָ7׿N^9|fz'3@n\*E?R+S8^mQ֡v0@Ql'^+~k*HR*8(I=Pܨ(X֪ VF(cW/g`~[&@)(&nL&MN+m'@5P&F:(-_{讚\H{mУz%J2R' G]ǸɕZٷFD+0ֳ%l <,k˭2=Ne邂sb#Ng7xceo/=&lE+~hcHݎ? s:E Z2y\ݶi ^[كoe9慖cȇC>Z>,{1,1.ow|Z)K\ɝDDySb{iݔ_]G/{wvegel>1H+FOwQZ_h{0l!wW 6-u.%un3#`dǤ9աbxu{4,OӳE⿝=v":ϣ+XҧTƂ{<9pGҍAks;uHTc׮jK\MW}jC_<̺+#yves}23_ Pſ6?ھ^;nJJEA=frLLF!luRXhjc ףjަ:4R5fх:"Bn3XrT2㎎V5wKc\gkkZxk-1 어vz{E\N?~|2vf|wkK͆Vi $ehԔ1bNµj/eBu؍q mq]y1׼.'ʅj!SRWkak#5>%U.D5#4$7Lj/08ݮ6cQר*EbR(>E:Q]|=u50sS"Dw"R5VYt7H!QSӠ׆R6*yˌwh|4gl"#R.i,ܢ^ PTl@t[ xi RsG" QN[@yU벫+ <.kkcnuEѕD]}"(4`kJS5#bf5+.M lp`W@ܙ|u\%6)@ZSH+՘;.̀j@o]\?,#7 c:om@1Kĝ{ŜEqN0t \"00QؗGcY\O` jX? `b@AVZ h\CMEwf%R9r pϋ55^v/J "F_R]Rמ3/O1 $3'I#52 -̒&dY2.2@b۽B6ф=r+`*Кe8M2ypr6/mb_X1o0ƄU86?jtD1!bE}o3߹a S9Wx\n `z3[eͻ ئG0{;|PmJ_6#XL"gHW}HY`0 8dփ|ƂBgs zjN҈Hd"UF5 l58tqXeXi /a\}l kЉ(R[q;2p /f,X_N5~d}y E*UG6Z pߩycM0 v=g}sy'KS=1ɘ}c6F4=Ro&x( 7DRD7$A#|CpB0=Sa4b,hD-hg*I+y 9z-| bhwXm56!XX-;pZ{f ΃pPDƢ pd.;ycm \g 0SkUX(ͮJPƁõj-Ƃy( !뮒}p4, )dlM'Jk?-BhϢ;4IyF&J[w jPʵrvlUrg(mUQ԰A6>O@zd 0L![[#h\7cȹ{JHu6tk*f쵦%'ՓFCon6L8GFo(̈=Xuk8n Ȁ-ƠC5q=<1C[u]kr%5N 0( dSbFzD@z[`}fAƮXD +IiMk ]#9)E bX0p1p#sc 5H,ܽİY;:@%E@*=6*RU 19f;#iz@iU{/A6L׉ dXR6UK!sOה=H#/z&D3aLMcVk ֫ݙЫڂi#U>.:jUae ja3 _XR Lt,KD(45,1\b$ ' ⼱ކ`ݢ>jʅrriDL0rW#Ф5|$#P]XtR@Ւ0[`4b/ ຎ C@rt)EL0@57GAӫ|G1QLkXx ]u(o;]q&vOf_L yz4m?ծ$`G |^~|Rv0.޾Jxs>o#=+_R?Ԗ[Z+ZNOVR'hq={ArQO\THэPo%4Ԣ96V*Uu+t2P]਴%S]F-BjGht;OsaY_/v5fO-MSEbP/:*FTT2L#XW1|T1֏$,z֋^mhE[/z֋^mhE[/z֋^mhE[/z֋^mhE[/z֋^mhE[/z֋^mhE[+0vO\ jJUcO&My [!U EH=Bc> <~eo=h:E`Aʹ)rIVR4Ns0;̥`a\+l[>]}۽Z6熂~Ohw77gWz_[m`lnmg~:]t5#KL+'/YTw/Z? W5hͽ~Zs?d>5 }vyVOVk :=g.bs:N6*OXR%k'l3in[겭|jc\꾄n #?\/cKٲul[sY j'ݨwU>Ts׮q?ːEپˆ,g>C%匤3rFRHI9#)g$匤3rFRHI9#)g$匤3rFRHI9#)g$匤3rFRHI9#)g$匤3rFRHI9#)g$匤3rFRHI9#)g~ ˇ{8ݳL>M^W釕gGbxk7_l뭢X&kߝ>KX)}@kKH|j~wm$G_!Krs6!EK aԒ{QDmNTUW_pV5n&:vk3!t?ݠ|dI>Waksuk"-Tq7^E9Z;߃t:vҌ.nz{0݂L~_?O`|%nRl}}eGL|LGcNFX۶SjBws?|JG"5"`hrSYP^?ͅxx6x( ݻu^Ιyۯ]F^m ?wZpԡQ'b IB!2d٨P4oQن.ejQ1^ԂK+ Zp1qv[8x4fͿ;bk6{^!&v!ز3eP]n|gDky;e@NuL-"gBR-#>h\7h°?O.LɕL%(aFh"KjbB4dT@=JT(e)(SަM8dX U\NF2OR6@l8ZQp8{ʭ.dbӥa@}_{=×[ -/ iNQDC acT,fI03h#$71@9МZ!^PP2j¸cs@k~V\DO-M;˓7mBkلp:z3wZ^_7*:9mwKΏRN& | Ǐz^ywFHTz&l.RXen\܄s_3(S3gg,4P_ BY}Q}R|ooivM5?~Il~\??bM*BIitDK"B&@4 :ۣ"D1,t\$1d"#{Yv\鐫0FBªA8= <]L;ںkvKgUxˁL%0B0(I<9_y2hd;f: i4x$j)` slnU7C 3]al˭PF^BMcjGTKXӳWP#l 7ajL O5zheB{/|ͧo$D pFQ45zL(%qCzwV<\NܼG$jʌmKHxB8+hFƕt]0WL8l{Bݹ9\bizO Ӱ 2_!bMӥҽ2^a*X`ZxǭN+m䖧<\HЦsz `yeb5";^NnZh;GWOQG:Bph'ιC:ۊob-_u6fͬFcыM""k цMFo.c JN1c_pr*e醙"0@0"jn)(n}tap%L.cC*[DC₡\1qƢ=8#JPNSv]yh\jW . dg ۩#͓B%[`:i$@ q%aG=uW-f F5u\<&q}&ϋIc8XG}s'ܢvHn8ScFz): ,3%\4dQ'vVL=^:ܦx)li|!'N3;$:jkÀ*%9mTD'8$L8* 4OEz?r(Yйux""d0sAF7@nae2a2tD)24e4g@cG9Xc'Uju1 z;kpn&Ҵ4 ɓ˥R@"?YlHMi 'ԑY\^JG,ҧޑ>KikG#}~޸Ŷ̼-_2SP F$!5ѤUPVw}3y]^. jT+1"ςRR!$v2w%7#j"MLR˅.".jMQ8>8sVq-Ո?|d%t5bf秴2FGrzw7巣$V9R_ASzA8sLC!*.$8MNr5h, iLkS05TN$c hJ5Z84P{l>i6(K%D_M80q G'_jSMM{#^ C N$΃S#;eH5tR@J3xeqL|<1$(=h:v0)lNm7>ks:# =2璿yFMaA%[,G<99'r sc@_S' I&7.\,̤ٟ?m?"V 3A'u fqt)n>凬G0f)UO#!\oۥwZW2Jdm~]\OoW?jbZRϧY$pqM7 SNM-WY]({h~vovz^_t?fG~l^?xmPsݗ0 W'ݓ]I ^`gAHHlq$!7t5 aVUaQS3h8xГ1閛C#.nkԎ{ynIB篣qȗJwNho4#?b6U.B ?݇|~Û}@78B.C%>BC۟ڊRCx mJJrǸ̇Mom?,gր8˛aKr]bu^#3sЕ0+(6lm?%Z.Oغaa@W^~-6-[1c;6ށJQX|LHlBd? hfCrJ# I :y'\ӉS h+%>)TaXd )*DŢ(W"x<Ɉ)Vv,eN!;/~{M獖⪺y\JA)iiجYvKs3pChtn b kB5YJe77IgxDta|[-OέQ(KFKϽ 2yKum1J?F SCZ\&g{edيL0 n@"ĩ\qT*P5@sQd1qDGcG_ ZZAxp>,d;/ctӕyXh{`"ZEeB$8 wB*nv~q :>*jVo h+y&cXfLE*ͭ2aF4&\pXR!1 Pk#t; J8PkBYBzf0eR .a,Bs&e';D),yy ,b zAMm[)}I7B~PEm%6MOP+ 'T`ĀFM>띱) D佖1NFSNz`m1M.s rChřdaӐƌ[k8ؿyURٗY8'do!0aZ[@WZ=,l{m\tNB9럛d]L6ލ}݇yitOClW$0`Yq[=%4XEI*"Wa'D"$Z޶;Ø8e*,g@#=1pdꝩU y&x8j$v$; p +4K+ ne<t\YMR#YhS PB(yO9& t#WI̓cԚ{r`3̎N4|!QlĈ#B1pd%a]H^#f)筵)w|J DG˗J;t{P0`}c95F8%A'{DI7*-k}Dosf,ׅX΀}_^n5Z?d&s&deͷM h:>ج6EGGI#\7!$feyY6zt*DWEϊJ&A4ۗҴ@ ಗXwXY$n <\o۪hؼaD=4T$K gvύBmu4^2 'G7!Vi R˥?`űű(7?NiW.[6M't^ .^9̭zv3L%䵮(F XA;׍ޯG-oe٦9nh+:BL_./FO_1S0ʝ[Яb6_WZdb}jTei *EhHiI&UPrWOeܚĝiۣ%9!i4侧!/Nc_S X9B&8aQ* bcfyt[Sf*U?3vNO, G~~Q8]Eupvޥ`%H7 2PCSP"UV{J\CbJFvB(ţV6jZn6(ѭ])xFe?Q1Dγ=4'nɤcR E!jJk%?PUcz-꿃D1KxcpƠ, 40(6,{NI$Q^b"P;5qu;cFY; ZDl hYj$bJc3tgM=[kSmi=+-+}5xA?a'dW$ U7Z0eae#fg>J\4[f,00ZnHZ]hBġЫcOm#gseC%/|0VhӋ᨞~Q%I|W~P|o{_,XNj+6zV WMk8s}?7r(vUR`΃e a13ooB/%9؊|eWش\&τznkX+(0 Jc0 /r)PmQ0bz2i%Gֻ1_+XW6nz=1w-c,0ULoU\P_W×_[wƁOߦY0iǢ3jL_/RR᠁}s}9"ߖtkrDKsRGX!u?jȏ_ڊ:5B )#"Ki7**@8#6D(żVHGJ' [giFDނԑ*m85Pq˗+8]g4qzwt.mdnԬ`]JF [16*ab- Z9乔i;Dp~* /z-ܓ` aͽ&hi1[ABAE \xl$a;=Xߍ/;r;K<&Kq ![_u~]X*UhcE;p<ؤ.̅hmt"$"}-$GG#~ @  @YX 9ׂ2bUoE+Rqp'`͖T nڔtA0v?^D Q=X"'`c&Y0C)s ǚ2΅u (%Ne A&c*5(K)"]h ݂/ȳi"aAtԌ^hZ@4|t`D1pPRp?BAFm1;fsvA4n$ɗh>ӋTeŚƓEd*#R"%1LNl*âҌ;$8#AyMJc&T6D>YlCFIW\{Usc1 (b <5I>AA w$䟲,vXKKiaDp Z(@ v-#ӺP BÜFi$)G nA`0 HN֨R5\[7º?c;7~k~撾$8?S_5#sN~:ID$RX4hT7NaoCI>}~x&(bwl+p"a9;RLݞNÙ=ܛ@ Sݯ*a@[i`OP& -,InFWsKؐ; FG¨,^|?6>G>dQRA[/ˢELJ>Dr x0;6`M  IcP!"jmJcqeE|\s]~qv2 .?x:<:.w.~H|t،[{F glg:k4Od4ˤЇ8(IGFOsφ5&j֋la,݇-f41v< S^H'qK@˪ @į%|\1|7ަ>_~!H߼>k:_̂ŰׄB ?okSSk0_=5UԪUW~!>.QzUaY3QϯGpR4K\ԋ- l~VN WX}j!'ݙ/ 4q?3\8ߥ&V>Z-q/(t>zB`cXmAn!i🜡(Hm$)VIaPFz";GOj753my 0Fa q$D]IĔ ^ϽYI,JSdwVwFI wm!IR[8 )%>ȢGv$,bKeZj˔-ֱRu7U!;^Eʥ-jaz&ˀeD? [&(l8ũe'_-Szio^r9_Jww7qFT oaA )MW;M:3n xvCٴckmu/.QodM"#'\h*4=@]MU3zIh)&8gj$-gE>Ch޴ePPgP'PL*ZzhRE`Uő&鳰IfL1oO&{`Bﮯ^f1c].V[?;k֬-ΟJSZ\\ǣtR>z텕N9+Si$9Xi(ZH"2 Z75b;meVuʫ }8 VD6ҡ4@9iU"ȘTc#n~#~@]\ \cS2>A{v2f_ g؜AC'Vۘli1KVLI)6^ uYY 25`=tJy:)1E郘"cD'O92eN;v@]GkրA#)KINR+ RG"5Hے8"k£(#Ch)b%eple,$Am3uFFK|2@a<*e\7(]U81 >+b '*^Y!j#z }{w7cܩ6]S vRRoCǣ;_-GϷs#a~Nw#;٭fxnh!yc0!oBk39fW[9UL|\ֳQ05ח:p xO1o_WGy0z_yoFnSlNbDfOսzyDC)RE#u}rz9Ϗ> A3R2'!x m6)L(8XڽǶxwl'_?gmǘ" ]<C~JJ&ɉ>kBiGu.Zǯ1JT)ix )R \m<|)-y,!xSeNʜf]#dqۥGϧkTӕ/huEoxzܘ\k6B9k#;v*/І9!4 m~as;̓t=b!+Ha+Hňш)YLT% 21%yZ#bsSirA;{oWbsB>TfPBR1mlJ6%;mP(1Y}3?A2XB?Z )"P: K|+/pՀW78\֝MdŒe&b FiV&V}C@ 4P&7r7~/Q.:XveH8,d"oT59 l%#y̟I 2V'}fgE~9.agbOH7$Rq"W7W30 [ 3Jo/WsD>A %=;nIzkUXY~>utKJ/K޳JZ1 r]\T?H9 mUr͐\0/y*GnJL?g~_.$\\_Gvxv]_y>^ji:"+Ƶзe1ߊŁu۞ /f=NgN)Ǜ!^(ݎZc7׾m ŃLbCX U˿J禶ƕf& X2.Wy'}#^CŻ1x#> ?+Y7DhYKANR`q)? `F7/E.xv@P%i6uaHBDRQ@ Mx04_]X ?U?4iwP19BM)rTCucZz# @"FPk L;N*AT$~}'W0Ņ'>0G/@3PnNx;n٢q6)vzv=ou!S%cn;:vZ!"0ő%'/APq!'TdOȿB 4 )B:sErg'm=8)K7ѣ$Yl1`JIT$m0JP4s˧YM|yq1}E%ۮ[o_/ݠ)A3w«V O!+,E0} >Uem))Zn N} >jz=5̊!ЂQhaK$p]P(K )>B=oI߯ݖмNq)=4?}9,ppM{8f>(oGoK)-ڱ)-u8F E!/:^ЕIX{LD'@ǂ d3}n1{>O|?Ӡ.@;e\)g&mA6zKn!H-(2t*#Pr [uS˺SmY I(ZJ O)V'(!zxN"Jim3umTZtE1]q}-P,IJ$2xSQ%FBoCwO<иÏ tvYt5kց\_O;{oPq;) ;A̠`c\?<ϟ3!U:c.K]4Pǂ :y ʎ{'B(1VgJ*_ZT6dbJJ$JHFM!IQ{/eH(FrU> os0(]&ccL9C:!pevs=8]бN>"(p~.P am :u8>:g/ff.@LlDFT(|&HHd ٷD Q%'e Fg%}TDTObκS.v:5:=Bahv^ɳ4>2};ARG>,nohÂ基ei Tz]k{ BM]بUC+(LБkА>Vs?NC ځ*6j#Yp&xJڤM'l ICLlӥHEgAHkt|vj)wv8}?|^ornn*Ͼ^Յ+ z:OggO?7};?C~l{v=:p1}cY5 $ތzb@~6|[Szcp9UmLnX'NPO'pR&KcWV 0:d\t]D]?'"Lo*AJN͇襤Ⱦ 2I}R3 )ȋ yS(N/Wݬ\onCOw4S1gg<slr2ēryI2)UN`ePՆ!+}1< zM-8i zR҆e$ S0) $' E\7VG6Dllo[qO`2Pއtׁ1Eh+"KA`Xb^4L.,ťhQU : !l G >2b (ŀ:9&}BPT&s@)5[˦80KX˙BQIfKY][8" he ڞ1E H"rȊcac0h*6Sgd@e 6}08>/ g`!Zѐ1[ "J΃EB6im+Al%$2k~ͯ2EHF{Ңue|+u겟n7gXJ2*k9fDG'ϿUǯǷ}OІWA{{&hNWޟQ8V:y+h!:<(A0ZQ˶ޕ T\d*o˅MBt? ;֞:=c;6BYC?*2|O[K]|_ծśۯ7Wo[6Sb;F!r 8tڣѐD{U. ++# ^PЈJ6u~)vc;vA%QGZc,oⵛ͎Ck'yZ Sb$tC%$x^yBf#) $-=_逊.&FGȌ e &Ǿ&HV2pJOTAdu'\8#Oqf˒*E! TEJyPg_G`CW#XaaRcc2e``"uvȃ f,Y4[[HZIJ%F˖l l6KO5a_5Ao9X}C mEkm#WEl'by9ss;4YHJ&b[-_d2mmH[_uq&DZ*7\|\<<!W{4>]u|9rj y]} xG R5#NJü"Z#oWU) k `Hn@k1RǢHg0FV<!F0)ȣ` /1R(J Ɉx Ev]]I87gj[wSɗ.#iv802I8_g/>!dDHlVf"gU$i,m&SLV1{{% QVK7nt_?eKٟ/UosDPu%cŒi}Vc3뤔gug]]Yc970MEૃWW[EnȻߠvmQ dz-*OiiicP,hd6XR܍`cHJ X"@m_ϋqy}4w_&dG#-1cH?GIKICB{|θ\0rr8 ~'tο..+'RV3J_ymFieO^@<^>˞Ϋzޜ~OSnzZCOȼT]+ܲEƧhЦiS4vHMP䳋A] tVx9Y$):n_]Gef0 `0.vWEZb)-4 Z33#ff9ME GRe )|^Lg/Eu2>K,;GFUu:~yuWKr]nQ.>֜ٵ]ql\vs&ф:6@>RX ɽ8Ydz9@O_K$^$! ]S3hUE'ԣ^[*Jyvjow `ZzvZenQ@xm! (yt_8( rC\_ <]H|7e k?׋-5kill,a 􏣗g b@\\ q% W+js\H--r5̠\{[Ճٽ->6<6zp(J WE\-WEZ<`ҰWo,2 jU i >h\_=c1%WC3>N\8>5GI)ؑ]# \9C?~wUzE-`/Ij/tAٖāU!9֭7lg{ۀJ \O*g >/\.i)7w Xt}B3y~B3Wjqrl?#3\??;:XwH( DVD(N w^t,^6JJ_闦D('&)ùJFϳe1딙À>FLRXe۶Z).T.HRdYp!A=aa2.WM6 %QCzpݜ4bkNtp]f{bAOթO/d:ԝkt]I;(R.yXNg,xչldWTC>|C>ܡd]("&`AGELh Hhd(20ǹX`H@#+Mke]B|iS|e0{.zl7/%^%bb"jn[ֶ[֧i}jZ֧i}jZWX)(8 ]=qJq)EJ- \ښVŴZVkjZVs ;B2 >@̺vi!X׫b]wkΐ\"/IYLA2Z#ī<&b"Ι+2"XƵ>ʱq:\ƽ9ot;/_VV&t˿ߕ bݍBVm07w M^.vŮ o~U}2DG^8!C@$Yt̆؟ak!-HC:<9"BYjPxD/T)(w[6xƐr e$h@eHYr19J:sap5e$V!&~i].fӯ^I=/fg_VBv?.{B7Eعϋ9k"/3rQ`pGABrxR!taP!dU ;\[K&+m|>cґQ\e &v g;M!K`e3`S1Lu1L$O%k[t(Iŏ̡$ơSe6VdɄ55}&6Qec"OgebL6Nh.5)4ȡ6mKrEf1Zy.W*2&Ύ僐}4ʰ ~Þd8lryOC%ˋiNސ&1jPXzixI1c:'ɠnQ`j5Y@!>ӽSLP"$d$FH%k'ʃvZl7q֗Y"Zf?YYx=v\_DOnl6u;ۓ톍ֺ Zlh.F:cKmRY /U{CJCAVRcD:GƏl4{5C ) IL > ,R-;RiĘ%')`@bPLH‚3)je_eXMXTj/X>+^ITƶ\~9g/;6]΍?g+Gl1e! ^er*)$C`BbLxRBa(3c(ΞI66A^+LF&H`u)`vZlGl?5XPvڦ2jڍ>YErQaDE Q`GŲe#;DJ<*!3Q 2šX" c*$s>N'd\=jlީ,sQX~싈X!bCć4DJhN2R [gS^F74K.JiT9""cI䖈 1=yI'͍$k I 0N28;J:M=[1uVӒpQ}`2.چ x!4kd{rG< k*X(+6!JԊ|Psa5yM<ޣ 쪈ך/y?Gn] 6 0e xG,O?BҠ @]kChJneD]@N%+ :*(,u)MbOíPa 2UZ-=kVjT(ա(֜ \F ;r$.X1r$pIJi`bɤ~I6Ϧcr$ujBw*+x4PkW 8W ;CU-~T8&}pR㨭b|K2͋(0*Ô[ΑJW \QՍ#oUʕU[Y__9S{e9Yl.f9i+=sM"'j2ABJƹqb(R1\'p8vl]yI[W`뽭=vjT֑ܲln,Jn~K[j;R\[S.nTԁ u2#\"zJDvy]c8;f;%d)JR1O6VDZ96ۤgN̘#rƣԀ6Dbu t2Mh O!jW6j29ITeM$ x6 S9CУ!&OʹIY&7dAmAhUz4ަ"OE7eæ3Ѝ|?UneR.gD$ Ԫ$dmb`%JrYfgs `IX >lB'ND!{7t@V3DzFH-ȉ#]$deR B`1%Z$UgG;+0^QRT: xdAfKHL4JGji9t$$,&8N* 4/EW@NXSLdT"+ό``2sAEW4rjOLK{FQ'wLal,;XD aiD띧 5(,sy*1O޵q$BM;Rq% 5~J$KRu|_Ґcj(Qvds55=UUL7ث葾DϢ-N+m57:`ذAk`bm&֒SM@K)9cDx"Ë>DdŇQ]^L4.r0.5 wwV!,/;*,$dՁ{^[ U; )!E(pu֤ҐH(MBBN6yd?٤IJaazϞP46: ' .XdB":1SM@YM2Mz̪GH7yVM$HEBMfK]$XlLSos,9gmcMZ: J `B12kQ%8H=Z†nJ}zcc5BΏ)ԌikYhn`7C&JIJч pCdEd\_5& |EEBm_.")TD' 6aQiƝVQXB^2(X y,7:g6Jj#TF[95I>a!&v6gݎ\ښ:#{iؤf4D0S"Zo8AN- HcbPIjBy!}ZS6;pOx4)G nA`0 hN֨RԵ;cc09"撾yz?x`oW Tۖb"a>*a=B d]`ָ4(-ĸ Yyq7\)OV/{أ}l)6K+pRx 0EdNΥSwbfpmnOfsP'J( ք{K 9 v-LНE$Ga4Oc,>,f`>p;RԳ3WME/_ί'\1u7l`Tb2 "5-MRn0X{zŒ2\{u]yͪqv3 ,_Q_T3KrfWE h_n%ڛ 3ZLUb |;1~nCodmVE=,CFRG YL;xtuLlyp`ޕwm-yTWnA|~O]"}3)"~O`\b[ GN.@ߟ}ûg^}ޜv ̃Km,(`~ ܹ5pT.V}>u=;"}>r},LW$thn _n>3oc':-LJz &Qr&Wt׾T! ,̀t%&Z >C;a1uGx!1h,l6\7@G4OPJ+0B(#=q(=vġyGXz,#7 5C( !6!h^#๷9+Ecr䃝^v*7Jr9yJ2r0q.I9L,*S}z{Zg ~9Mj3Q0DMqt=EctEG銎U{ǘ%<`Sop3rƠ, 40 Dp%MDyI~yRNqpcp3VA~ro,SZ0 ^wm,zh[m. \Xy x˴/@[ka=6O$#?ѬNo7Z ʪ'W@{jHH ;{ǣʋϺV"/l0`:$ M)(@ "2[z}_J gX`ӵ;X^l98C.|kڥ`87S)CVR& 1R7ؾ0IH ۂQmQ'(߇߯I8g~660OH; SVgD\.$Be f#-e8JcJea/9=Ha9cH; #8(P1:MQQR!ј0ZzR<UXmyM a,`ZFL`8MV$ў#7#=Oo˼Ö0!iHvRƌg[kܴvEԖ<\ٝy%d.ܝx2/iBæ+[Y#2Ϥdm'^ lcF0|4Fo'9 lad#ZZ6w(/&^7@yevy[5W-̏G[qkeٷ퀁ޒ]s[]>1.h9c猤e|s-4#|GSU WH0j .`ΤT^I=oSG穓m gxȄՄUr\Ԧ)xZF$aAp@Tpqh)Sh8K@@'fv˺ŗnh>>^xc"BZV?SB51:UZP*dpd:e-)1qE4 TX@FN{ncԀJ LqZI d156L<$I(RRaqR"!,s_Hj,qT)1HFƸP&hi5J! iZ ;張9Exo;whx"Fd_&*ɞ0W7ǖz^z (5(10J%$Mitx7-#.ɋ߷7Mx7b'>;l!e tdñ7xQăH$ȐR"uLɓTY)q)[sd ъ1#F'`^/J/IQM6jñ^~6.VXqrNauQr&?ݠuVnk, __YeUu-|PPx@ `JM[ Ŵõ%0 [>A}v[~Uih */SݫᨹA󢚫Iy3f} 7AV8˞0vd8LN}6ş_WPe#i@RjTVdf4r,4f3q&2Dƙ8g"r3q&bȊ!Ld2Dƙ8g"Ld͓q&2Dƙ8g"Ld3q&2Dƙ8g"Ld3q&2Dƙ8g"Ld3q&2Dƙ8g"Ld3q&2Dƙ8g"L7h{tAgin)rC4vԄy] QS(!b;bE<H@y>=g {1Kx SV018cP@A xDX= (/H:RNq,z̝vX-"ro0Cɒ2Bs4lNdR/kדT[y >OM/>~lK|c6Zvu݀)5#&۫OdrUgdj^\ESSM +/>>p}8E"_`luHXSQTAD6[d*֭ >jsRl60OH.; S*5 IP DقH kY(0g(,()9:(Rz,=y?΄θ%"TDE06F  6=Jӱ|"63FKvB(:o1c2b=hj4BZ"zNGޘ(ھoh>m6XmOnwWT歵d]@pGy_fuY;h&ôfwwǓyINwJw&o0x;qWLL2v':P`w1x 4FX\-߿u6Uۦ> +#t:"Ax dκK* [Y$mQQR:Q/97OI9) @$}6e Sq&1(m&fya H=E>ճzñM(0& "<-V]?غ9.od W`0/zw_VO-}L.j߬ +p2' e:`SAA.:Qg4ID| 4O(S3Jd# 0A%+er \1TDPKm9CMTkRʢhNV5MbuR"(s.+Nm6m:$Fivc[/ _V8I{ sC7ב͛ǯf7lߌ7OgRZp;9܀x22FE ١SIkHQCrq/!f X@ڒ$h a5NE- I)5Ɓ6a ~ /;|:o?0>-UNM1PgȽy8Oy*MyQ./uZdĄ|g[Į(Rȝ 09y y Yi*g$s#\KdQ6EٯRnk"nrBW$i49GOh:?S_h4i^m~qo=0=8R6}S}[ruyof*4.BdkF7ݵmY[@Ri!qSl|oM9a\BgPyRC0éPd`"*&V*筱t "!Ȼ6~ۤ/cavkX{Uu ηR16RzSfgt qѲb *)1%kCJN$KӮ-ypAy(A[]_UX,37H 2f8yG*/a!l(9)D;7OQ֟zn5x/^np*.drœ7ہi=rB.i&bY}Vn9k@EiBt] .՜}u]Q{TђX׫b]_ ֋'\m.%p@l%i(N9hk *dID1t)Rc3jiY(yZm֗ h${ҍ쯎єc:LxPrٮS{&%:ttšt3_%;z~Y$l:TVTlwvBbYpxCАti@'%"ONP E碲Ec} q>E׶޺_.1͏2FDYI Ρ4QhpuaVӹ_Q/Ǵ.n7z~n=t$rrXWl:Խ|q&DH@ "JHtN9Bf-T;4tۜ~$mlHil!33PL@&HFr0qjJsmfب&Z 7x cWDt 8 ]o!Z1zV*"e6EFJ)B誦f]OXȊ'ȶ1i陘Hy0ق$C5-%B6Ffy@[g҇l&%;~+Dc\.x#d-o2sb M Фs=̡ꄪQ+IͤcW<-p}_{uS54iG |௪]ApC3E?*}kjT577xr-(&L0d찠<;]IP[)9$p1$^st=e%S5iKqm A(*Dt ʑgRTV,JKhT&&Nik%OF2</BM6\p=4wXK"ͦh͵Ld_wAd:ZfPCTl=~r[n¾w1sIThWYIC:a),FymW) ~M.M.g52zoJBdeN8:ygX@M*d_ &  1M xga)|ne}Pj,bvgNgD`I"BI6ķrμ ѽGgiښD6mte<@E&ΔViH2e A`"`6E< 1BQNI $i:Inv*|/vh5?kK䑔N+5Zx9䅳_ٞ~9cvA.dŐ>7UuDzaf.|eJ}*?~@௣[dxstlTPRH\[ !ٌNl0xzD-_[,[%yV$@5D$zѡn 2N{}#4Wg!Sє>7 e2>>Y,(bAt9; NhbHjlH7tm}gs]T~f54=1WbϦO]9>8V*q6kxu5꬞AJCg~ѫGbn怊ènF!KL1^J'Q,' qm%:U_kNB8!4붷jg^JӧK)l[K ")H]4bB ؆s2a wmSݾkֵt,Y+P!u[b?Pk_-vMSf 3ٮ9Ѿ83\Bnk',&wEBʖWɮ$_CC˵q(ێs&rL^/4ʂ9@ 1#-B Z1`Χ s*m[L m2ZDI) *PuN+<"dF ňMC J2.MiW'ʏ8~19+o[w B ^HJ`L zM@NŘm1^k$ LsWק)9E H2] MJ%!F7"6ճ J@/$ 2IR^9 0)] qD DXϚMzvx'S;_>DΌf=&ހռ3}0 8f( E *R~-3?eVzв{rsB9$bC&221`NmaZVOI`'=/;@Oe}'Xz2ƄU9 ?{ב\ʀ_z~`M:ʇ 6cdp,{g#!l5fTwU:nc(Uf5 /ԁcxPz=)6IL`W>z Tdpֳ25Z+<WD`[khsFl!r#1om%=y9nu6µfOPG.lL==;mײ*µ[ɏge nۍLY!4k0^ˇ', 3h6f(hH^%}jd/wJ%;!gZv'85/gJ>r4ېsȴ/:x|~3J8gGG [Eb~C"_=:I 挼vN#הeAJ|{twOx`Z>:La h08t"x>eG0WZ}{9ٹ~Qw1x6|jcb #Bpc& xR2͐iʹ|}IE\l=*:xSN[kVãaNp؈{k@:kt&zS\?r)vW}䄸:K(of0%Jjو?^m]iH'~]NIsΰS{'CWU}pvh?b"P\{ꦭ!VqCU]}W 4G\Yy|X0-y0pveϹ~mVO8smxB mcd?eo3} RRPJS9pWL( Ee(2mvܸj/x2I=Htix}׬?rV_t\ޭ( ᩚuq5Zح>ٶS!LA^ M=(rם~hےYGKB I)@mNeb2e%%CZMHwg?[^SM,$n_ﰳޑPsZkK B.:dxIŜ OZFB9;򤵘ZWz@n\RMR)G,YV.0XE0).V{.CzDh){eX 0nRB4cVNcnTL܎ӤN8_S AiÄOD4!rY7$*ybȔ-Z(b*ۖR!dFt$X* @2OЎ*V!lMڠ 'KJm oA=LĬ {Am8\T>pօD2Hk2d>"#Pҷ \'SNOMچb7'MU(j#sHTL9QPJR o\{YG$+f7xS<b'P(JW<re_̵|DCZ4Vj()FJ[KIGR\L[&f zXRŤRRNl9RTaK P>@憠FkE„]TerB#-m@-d /d@X8pi-V%S^G J@KJZ 7!yA3 vX_zVc!# yPr  2QӲF Y`vQR0icŁS ^\D,/((nU;v`Ee06 ی2YЩ֏/E>ONXm l dNIC$Jb)t[7YfQ9TD`b"WCj eD%V$X@H;K!z) z1΁6D6y-> 5 &Cb,CJJpMކEAJA3uAN ϠB.q9NisghX%Udk ̓;$@(AƮ촱63+HI$JI>2@~ЃA2AC("mXfh%b,t0"UcD8P!YԢ AN$N fc͉L[B:b:K4AXF$Rhě*CJuκ[oY^PVIcA-FUP,'ࢠkvIԪ0LCKV`42f Nn93) >(%8"TSREc*K3kϻjyٖaL gfξL;**CJ6VPX?k,tÀe23^ N F9@POjOYSS9`8e{ %)𤖀[ 2?NB<`\( ᧍Yc6VŪR盉H0IJK1:Fa l$\E!*LC$(`r * rKSkt!w:;cgP$ZʮjB:'WƓ-(Jnd&mcVգN!N߾yytr ڄeKN0$gUN[m5N o֡$Mt ;z'P@wN ;z'P@wN ;z'P@wN ;z'P@wN ;z'P@wN ;z'P@EQ' F>Nf. ԬmVjVZ;N 9wN ;z'P@wN ;z'P@wN ;z'P@wN ;z'P@wN ;z'P@wN ;z'Pb;6?N l?ػ6d*qdľ_ qN$ bl吊`TQjD,jñ"Nzv%y6H VG'@/  we$PFe$PFe$PFe$PFe$PFe$PFe$PFe$PFe$PFe$PFe$PFe$PFe$PFe$PFe$PFe$PFe$PFe$PFe$PFe$PFe$PFe$PFe$ Ft@ ~;?4t.iE"W2#2(#2(#2(#2(#2(#2(#2(#2(#2(#2(#2(#2(#2(#2(#2(#2(#2(#2(#2(#2(#2(#2(#2!n $Kr{QJ5Mfr}PہdGmveK<,Ɗl0LQT@zy;TIddfԳ 2bΪ<˞=p\_o^v)R~`G W>@<̊Oʃ#(LqQ9@@BqXt~8`z]· ^* Qa'ϧ.<4V̦aǸDqCgH)+My=rB{djL] Yhha|*mHQ-Ѣ/)րGYdهlNa6Q0d 1*!xt*DT.o5<2YL uI *A=h1y7I^Bc-j ZrV8Db:poOjKtnaCT&Lo^ܾye&Ϳ6Fpo:&LYDŽ˵#ji]C_{tk@|_,`wCqCT?3u>Ӱ'|KgLCJ58ty:aN͇{k~ki:7AƾR_п[q?rߏ2ĭ=G7Ƶ==/z$:L>5a=i_P~mla,:p#:cH[䥴itN J1B ?]ٚ>1;I(ҼJ+ D. Ġ\Xp'h)5i-ٶ\T}u>kB[Ue1)V郟bEg!er5N\k WT mC'"6O&}eVKe7(֤inuvDrlǎ"|y $h(4: 53]_g=GTLE->%ȬQ: "h I)6! y%;ճ.;Գig^Y%} ?.Qb NZSTrXs`r!"@8DfR u ;|:ω);4"=|C($\HY‚%a)r1?qoݏq9u&'XL +:VS3βh޹ ڧ*9Bc`k)7/~H‚85X(tʂptb80Oh6#]vcY|2>%\ ~oMuލ/'P6 nUf=J N%N;bq"Ucit` 1 OΕR朵"5eRwwPrK+vo1m5{?>7L f6(RVR >% dùcF:g !jDs6b:շjQ0w_} <"}#U6lzŵG,73ʩ9"sjV;>a!"$$Y?jKKi8OzÉ*pjYF5HR#ʊԭ"i]@\DL nwOQ ,# d08,oT)-w?k r5+?W{.'+NHo|0 n|waFW; 7 fl=]6uCQMQjh\(x-|<^/_tsp9(Qcu9Ȧ^yغ24,5dQjxZa]%XK;}|{iX" @TLcAl0Z</o{Bۻ߽U/fpĂfᗇ1ݏ֬eT6].G];.-tCҫvV+k@~z; J7ǠO:tZz!OWn.l_.ۃjWns%ZIw+8 hrY:ڈ1H7}qV1PKkFnC! "$ 3T>Z+dEsw67Ke:tUfH#!3D!ē w $Sxx+m :e.r0q‚N$SQSFMsFV;|rےr z`fV>60O(68Y(% IP DقH kY(0g(,לwdr%J8טXD@љƨT?J >}yY G \-zgl*&ye4z hnE*<|g :XW2zc4^W9 ɳAd׫`AC0f\.gacWv5Ԟ.wW4cZ/wWǓYE<\ֳj^ʱbb#tQ'bvO<кSbz?yx6w6+вE-{(Mv卖a:|?y~=62?v9;_pq#ck{p+\ٴ-q抛m>Ghc&Vߌrlb[ ˝`q];Z~ seW".NIg%\!!&(bPB`-cR*$/tFyV0L13{oXMZ= GB:/<6M1kF8 uGrcdMW1vx,q2[/Zޟu G3wRƂv`, Anku~"k"1:UZP*d(!] K2 c⌋i#.?bV3Qk%)|`3Flv[Ny'4MU hWSVhqqD3Gi!K>ь . (Gb=嘀܎Z% htP@t|G 6 Po(]`PMr]on3]?r$~ U<*:]ճiHr! ~;iK`o_)yzAY?>WY0~)X6&S ,(/4Gu}>[x{q}~IO e0}2A &+DZ(ֹb ]HJocYUS^ؖZ$X)3/ZG[lo2 'wDw%*M͚G"jDm#m}|{g</y%ꣵY6>y<[@&i^|xPFZ(XT()ɤ58"k[e([,[ժ[5.jt\>Cw7PSfj~,smf~:/1.Yh9bBRꔪP'0N*Zh 8TM%8yKt1x^}^P}="X\`V1*PCc n֜ x:os_| [V7E4W,5]jD/,BF}vqafhq5 iD>|P(ERJtt e^&$[ tImjN`EۜD6ٶ1 R_'Vx0(d9{[D̡WThl1m֜G*[!`Z_HKA{c|0!Cjc629yr>JA[4M? ` 77QoQ@(" J+@sUsv *Տؔh)Nk0˧ڀ%m6Ol,+};hN"l;kwVS{CO))AnFc);|^h w>XHTۜbJ7ΖKRrQ`m.$}gj,hPZ#cFv\6bizg/.4xCޜ*)5y0MGg/MD9jQ2LDJUċB_ 9fmé U YrT¦26Z7[IQH#(ޔ'Zh՜݈'|K<n6;EmZ4%xwFDE$@90s5/ZP"GHQ'a:D*ؐHC&f(Zš<%C*C5?R6n֜x+RN/"ƈDqیY`-='-H%0r *ZzO.ki%Qk[ ΍ҍQP1™#ihHi VPcZ5g7"^" u$\J, +i~xP)Y ǹ,WxK*Ҩsݢ̡ʄ^+8I͎}P>v{_԰ȝ/bKv(^W~Jjl O"S͌s D\6(CfsToAWAܳ}mpjfOFc~sjc~(@-g"s@b>GK9!vf/DZSM 071)VzDKA W9(ź\5n֜G+`|׸lǵ%ggM)JWw?q@R֏I$ RBE֠./^<"2YQh1FƄOP!!8Di&KHO1&"DQ:]v.ds,Co|#fiCA-lUB4+H{N`sptaya!Udt`+氽Ni Z!qn  -r\H`jhOv" O~9M&^&Pґ|K~qvz.wOgm0J95W"\6`u`EV#+%3VKF/!!JHmK8~MhACӝiLͳ!Gՙף/,GtTC`0ϝA9~S~oBңa6s8Jf9q5nt @1^w1> S\ZXw}LzyJVIr-'¥/Qm2cid =j=|nQ\]=ߪ2CrNi1 ƝÕ=fIa@&O(?Fa4ʏp&M E e^8Hso8%"t8kʈRj4KN]׿u컥|LJcHujN#nj)gD88MOf)Z!j՜Gl5r(L Hg\F &)BB(PBpTӘ N֏gمniEWcv ?\2pvuQX\V>񏎍Y<cZb}7ݻNpE~X}ߺLJxo?v7uݛ-y{3KYM}9K%-"\/oc}En7+; ${N=uCxd& [nT3U!K{^,Wũnu;9{*]tNuB#2ID]E`sRV^mt6'r;f'Szr5zBlbyj}K)쏣ַb h߯4wt=!AϯW=xGU~FywltYm/ GV \H9GD*lqh6'zq#ϖVR{-k>rIE94\?4;K[ _kKql!/k8넦G^iMsH\KYH~V7k޾}@n<ʅq^*VnRK0`l ({9stՏN1%)[teA9Wm4vk pl>OrKjXB1+`,BTc՞ZpdҒ/h1\躄ӊr+ȬRۿu)kLV &`MF%Uz1QwfI` 6ƨ: & 3+]!҂Wdc*ژ&k;7u؆-PodVemgROgԗ}lu\}c\Ɗ1OeU}E^NCu٨IB0s7$6  [Ƨh|`d=A#2_,~)=:1$ԶON[=%Udʤa@(^ws *^{!ߪ٩ʚ ŋ9 qށͮ"!PiU;( =]4AnI-V(xO?UKM&j}k>|!5~ ^6x/egm clm ^6x/egM92m^6x/elR=RcePjllm ^6x/elo!qF),X/}nFFE:MEY;ȷH*_AQ%/Ni.FZa{4@]o u]WWkZ%J1p> (\$bl)WXlV]%4xc@[Z}u{0k}{sҥfف4AԸo )G!Ƥ[GSOlBĨjV7s/qq㣃nWk_-;@bD/Yz%~d&wNR9 hlbZ솏= \AmOQ)JhR7^k (;aH1ePV&sqD.VW&)Bc)ɺ_/Eč%,:mDsMUDaT+>Hwn3eo 6=7 ߈ l+5 & L&"-^ap^~*Ct#"3.cox̎HXoCpBT!,%݆ KhyZ6H:|l[nz+Q;<"5?>Kߒ+_<#;{gRK,Vf㯅WG8B!j9QzbDBLspFekۨd*}}QAZJ%R B(jAXsAT軈.!`S'3vnt4㡾:B?|sڽ4dfe:O|jr?=aJت$•*8B^"!I %Dl˪̬!/Hch^*FY"cMs0#p*8*]7H/qt0/BMݴ^;{m^{ 4ާ) !*Efʨ]p6EU6Vl]&5pz)x$Bd[-*a2pJDѤg S5nAjxsC=b#ަX4$JPdL@o8g%bm=-f؄Q O)@'1{gu`mA'C AkI6)n0M{4eIb:3MuvӒEN~18mlΈN&~Ji`1QJ$> Y6-Ql´rJQuŧiCϯ \X8rǼ#e\,8znT?PckݫVK3҉Z4(Tk)M0h5`Ԣ7ʷaUR@Y jE;--(8NWWbUFDA+ ߀ 1Xu2+⦐0CP4rH$Tа"4yG.Z-RnܣI|&ѽ.to2U^KB,oD$w+r\&/qVD\dgNXt*il Q { NvXn7˝ao΂Ԓ ղͺ*͋QpbLbDt&ΒɢҢȬk2FBΦTU)vM{Ү~fTԃ)` *ju+l#X2egE8gUlӸEh@O~am'V&c^W%ߞ eA0Xʾփ}JHPrU_=A|<$7#NzY ҘK ж0j39:SP3V"% #x[k2QDBqtq >>9ܔen:idݏw#{EjTHѧZ&l!,QcRDX*Nxj@Eon*@Ӫ]?yD_&b WSHe0ri/YYS4LY1nG?8c<:tg_cd&֜fqiyʴJҵ(K PV 䅊M@E׿dUjh PT3)ѓOM5۪WQM'?e2eݮ:Иl&<J)$L˙ r:Z}YP^a1 m²GDȨ7%һƺ8luM7)Ūڅ샏;k˄'2|Q$T}J*Z+g,f(%KsJ\U=dlZbS,Z+mjdۚmK$WCt>ǀEpC`ssAwSOop_jK&8%lY׽SEK%Xc|񉴳D6Ռ|}ʔc|Il?ii?ty$E tjmR,W P$o 넉8MZbA CkHQfjk cK!ʘjd^͞Q*iI<' 5Q{;q=aP~Ͽ:5Onk1[?zNf4 ['٪>+gJvR3aw-\O6,.U,,5PǛiju1NhL[n[ ZZFUfPmDVۼ ާ٧űmݪ9yf7O@.Mx~~<:?HVښOi0-]S'_g[wvw.5y0$e;$pun'2W&G>C8quzzKɟ4?N|</ :|~鷽y}=}6~~$hOLwx'Ԛ$f\;7n~k*?,}<L~0}Z=1o91u\7{um󨋶ѓW>=8' ^uew:8 z͋ :ۛ{|m{?ͷ?|odZm)Q^]]VY}D3_Fm#]b)_J:|y,]H0n۶ %MJMHIi;q):QUOr( V O9Y)i% !;rydeL^*F @q1Z@H\|!pАjSftY٨n:FJ_tt+عVҝ) j۽ttfɟ˯,2Ҝ'lLi,t(q5B4S+0lߒp=E~p jAE)fm%/{WFJA/kLIyNgӘnˢSR&5$mY;UEXdLJUlѪdTeơxU`KNg 7WU&{G[R~lVkRҨN1~~4)(x)H'6Cלݣ}vRIՔ`ux4=bY9N5S]Om^Q2zY2 CRaafy`omL2e*j >!ȼ-,õ+P<ǝǺsx Ky*=%.y {!Z1F}Hx% #ˢxQU>lS>Y"]xEojvGI?S4>3%])3;ƙ_Kٍ0/A!jJ%kdy>$舒lHecp ~1(  <",SIX$O٬&`'yS(sVA-m=KقCɒ2Bs4f6qu3~ Xd^ӅԥYx4sm:~j50Zfuـ%#W^ *lmfM2+/y}8E"_`luHXSQTQ'W6JyaIm֋R&e7{3qd,+ ն?Y˝BW3z,\HNqhsu9{L,c02}SaC'%GJĨ6xΨjg5 g#c eDn2h{2%bsez5"5bbގW֙aaC]Iؤ00axhfKd43dk9Uy3rq5,%]^kq _!ppE@##*Ϲ`1X9 ^RRx\]t_DG-*" C<%m\Lh7ijCo(<`M*E!W*8ke^ԠX=8)|^K"Lpo?>+սVOG%S mZVo .wnv^M99;YFYϐ֞iT `⢓|TqK&VI Z"' JCy*9MLt`j63@e5L+Š?!S|?aNS./0K; 1y{ZӂnfyAܖmErZ#j'UW-h4";ڏf ZpsЃJWϏNǗw'Xc)h[ܘ$NҭXXY/ok₻= 3>ç=-`8j00Hśޞ7yIÏ'J䗻#~O~4SD<"btӨ= A@Z#D§?B\;)IhoOPF3uR_B;aBw >EO{t]{`k%Fk߭7߻%\klJa^i°S$#Y@SJx->c}u`[r**}V˰]eRaXL]2nPa+-]k[zm> aK QB׳D+`04ׂfn-k6ڥˇmu>겻#`.{[]]1z뻲2֋7 %Z&JZGrH]1@1gz N BV[!}fV{= ~{![i~jfz^.=?󮙶EӮߛzd1LymIm2`h|89 j EVb%#A:QKO"z&GWPK}/Q~ e"%UbHBPh5j7o l{̳ōG(U{ݘE Ͷ[im3|iL/]A?˱~LNb E 8w. c"0A`=QBJӘG"h@QAsglrD`Ij@C&gv ^kg}r<%ًםaKLQ Gτ,Z*C44fPe&R] ؈ӡdBђIK(}4 ''5J~%rG8eA-"Zz%'qp6r}Б8EPRp/2ZJ挵30{-#N6`+R֝r>t.G࿡?Ou$k!Аqh䔶GR1#1$c"UPC7 ]QQ O>.0 B" T)J,ٺ]PCUYƩ}π%'/=1p:]m.<9"հ֚>*id-%>)jj'jsi6y:8¯ ^jlHi#S G T#Hͪh-JTB)) JA}{,32x5sf"(ܚ1[wk|J6]3vՅ"..FXCQHPDr=@& @2#̡R$ dgT`3 j-$M*)K*#AEt!"'Uf;5 <;v2֖-؍ŌijuIA)FQ#9:%YpzOY1 !N@ A ('p J"c}DR M;lٮ+P?+jDY#^#qKC1T0Q;YꭡV84:IG&*o]6Q0K- 9@҄Ij$>[rE՝2R[ЋYc:%;"BYb7q{>< 0lu74E\d7zg~-5ؔ=gmJfLUa/K/6Pv9s|IwA;hi͐ڍZcwW rhSʐQ6H ``d9{c06\טE쩊0Q:`Ly$L)܎Y>%w)u!^#֝gzȑ_m`r3gY &{/mmd+xjɲeNg.Y]UXgʛLmʯmH2M.mj}C?~7]$ l @/RRi6f́tG^"b *Bޢ, 'd)'cwpf,)dmTѤmRrJbC΋, yC *rsԌyc 㚉 7Ka'xsF|JFhN >rf83RIQhLBkNCaOQU`OϘLAJFr]@W?[ּf(]>1L9A3ȷ4>G<_Bp%4FzA =97Wg<}0)r9ZHu樂o4ߡP G ٕeVSJ-8{@Z|uj}F\($zkY':*zA FQuJFfcJ`f R9JZp\ʓJbg/Gq3qNOL 5Jx@]{F[rwė[jch|e3wt5A`0H grf$2UaEihFBtz6ۉ6ky*hD'44 3IRײV™䍆*[b2)~Ue3`28E IHf,ro~Vx4N(ȌޞIF-`>Kq *iefVeeDjujԟ8y Ke1H;8 !$yt'Rea]& DRhpqp~s}*:ӉSx/PtS!Єm>jS xU)G19sfDH)Ve^w"no?xp*9o+R @k|9N`W \;K&(?U?V38, Gid ~p3.J.1ŌS$_2cn?ǏQ1ٵJa6}lX^.,0x6%3} (iƁ`욝]-xg3 ۳ٜ\\[[9:2ve"3ԂY7.ŇG8'L?:Guo׾v%S'oM/G(|>}р޷ݘ\;5oh,M?܅RH~pփE^_4կ+?/ޮg[`m8Σ wd$ho5K?^=,k0cHRmm>5ì;2ÄI Xfxt|г1W##jݣ.omԖ{enYHK )+R&Xʃ >3L;-bSvtI'op|/?~_͇?~|C 1>6x>NNv54ZXж˻nc[n 5Z?f2-GYl3 KU8 +$6UJGlaS*OƫTKaa@2ek{qH$yDt!Q'u!E:dpg2$G)zи` /%h&8Kw;N>𩖾0N2B֞$1{s҄#'Ie VS_d]J_;ӉWRz R%KwtGrT;Oj-Z$5zXZX]H)Q+靨$J1geV$-\(X_rɧN`'vd&+ DPQ_SPj, d",۝\ Ia$,> { i%PF td\Qdg쉎4N~=8r eaX neۘ~Q\{O8|ttX}֚*ߑ}:],.j_Vʺ@1/uiM@ԿQ.NGmhp_3>3- < yIPKS99.gL8芮c֓)(,dȖ320LRiWGrJ:D,fmL!-Wj?%)R9g/U&g"3'sw+8H鲋{N4Ӟ TNF8뭾3?NkcZGC ꫂ0<,g!;pŀF]ԝ|U.̓I Q1+]bš'< Bȣ1z7~?{QSsVrWQi 4' ) Di]U{r'`\XHmQd4?ٟy:zB:x^]Ld8g_}[M2YSkK?GuzsGח$zmr͆Fe9uݨNjF5ESs8;r.85ɏT/O g? T/ \O5G-9I]ýs^w{Oӽ`_ecj~?|SmQY֚"ks]\nDۭmۻΛfw->oA%6jr-= iֻ}|ˏJ)lqrnܵ$U7`f>]Yҭ>PD힑#ROmDZM#jYnhy@{}c/ժ3jIӄ4; d=,j6}]d ȡo+ą&"cKОIB C@$I.2 aiwlCS"9,Ҳ:HJ"%,IM=HTU1*/2`)\TQ[v6=đߞ'NKQ-gpEb$kE%t4߫r7Vnv^f½U_qI@fHR[@.}^ќ\u˹q&I إAE5~!ƹ#ZǏy޻hjHcp 1(  4A% U#$reqc*h%{gu(!)#4'(xAcl^}2,87Pm?7V.ݾۦkhRk ˳srHhwJn bx*Q1Zt?&Y)r3q %7ih 5(o+L )KNn[am^۹gP#QGkhNQ#T!"CTD@cZZ=M[QDjmr 8M(rne垗, C[HllS}8Dc[5~|vS P@mmx}".(Cw❛ `xV/LLe&Z>~XTs@F!:?ѷutB[pǹAqf>rk2'UCHK)@8Ա= >)2O2sf(:SW;ie;>)=[ҩg;ӽg:ݓT(7/\Ys9?4<0smX^ V0R[l7zTnٓ[ô5Hs.gHI +2F %G%À 3 \kuW `k U cJ#6 OL1/ Jx-j3wZka7ȍzO0ꔪOuMABXdܧ_^gwq)͚Nl*,eSb4MK4R]ʫ`B:u~N,d 8E"`DbHkbMn4FQ"O'RT,鬫T\ j~JKՔ}}tc؆\@|ҥk66`ͲGixmfzzÓۙO:(rLșb:WڢXj>z%qve.v,MtYYiRYS=}RK ޠG. |ezlekc,}˺?@>PY>>Qoq8sEck35Ȯ#et!/K% ^Ug k]Sew[Vab3Ézneţ?=~,_;{=x|cFP;åu א?M q@D^^?F.`dܚ@Mή 4;TA.SY<5WIEfT'toh5`@̄͐5#,kB*r(I{NkF8 unˆ^FW2YT ҄EL ` Xy$RD[>Xn9M%`~*Cէ~i6%Oӊ_.+Z=q,c"BZ)"B NlOB(rQ4B嚍A1@θhF 0e%*^eq2uni~:q4MUy$Ѝd1fPw|c_qJ8*Rc`|fpǖoMZ~KB]!+ Uݽ:֞$eRT̼Κm>yE ]y)Cbp`*?,%˽-WAMb[NQ=C4ggĽeXU}OntG^Obi4UM<<ڊXIKZ'~syCT`%%$O˷JtcWNyrP`NT4 } .5;'>ϙvrh e fW­=y# ;}, 3Vۀhc7`Q 2bnGS0<T,&  YWJ !8X%1F/ hiji5t6H:sNĚSVd! O"TIS*xƢ4-ͺ'-'dm?bx?~]m`370~dqt;r`1 y+hp1yh2*ӣ͙s!8[6Q{BBkbqsz70b?[α]nv~--/ݹ}>G-ͯxa=WsgwVKOj;Y?5?u{㻪 ]Ҟ3_)3sW_fn=7#nne`7gIKBdɈȪ sZu:(NڇL}Ԙlf3[Vv29t0FZMuRI%'ДlG^| xVY&L^YELs\HÉ܁v42P.$#@&h qLT"%r ;ңYq({QJmCe0[dvP99/huQ8pYkkd2`LF?<෿ h0n 0N`?}xևg >Cɇ/9fw)YNKy*W>hMFƱq wtycCEC Iś3"9ܞf濥lHhI]\F ׋GfW /~=%<-k~[0_G’?J~؏9\h-t_uvCBXɟfMX~C{3irP |&bq7bӞʄh0 QhPmThjSEqT8_r)Mm[ohx*˚^UY=lgA ֥8JdIGhwHdžLmbY^Z4O҆9Ƙ; EXЉུFRnpe8X{ErQi'Iɓ: !u2c+o&EI`c۬;{3}pNfωx%,YaBkL6}95 KZJI"tYTB} lB733inbA[Q6DW3sNq̉;l xS(zm@w i`&<B>%[]L@CgL@N6bh%KJ ft1N(eHbw-K/c'$ccyOw/k?¦qnڪy[vݽ7Jy|woR Bȴ/ 6 :s y1.04&|WBk/RϗCE/{>Z;d\EѡAiRڂ 1GW*"r< Te "QdK !gHD-˺G*Ь;  *hf=^9mٔ=0̿k#:gJEf/@du )#ÀB?KC@"Zȼ6$2/+D V E:WL.ċXUqM'j QιX$UiݵYwvCZ 9 m|>Kv}= ܿ _{FGx9I4i5BlDʂPɨ5G#>T݇ejXvH.I :$ML4Dǃކ\ȃGMP ih-F ^_QUH)e*!ALŝGp1!OYL;g6s)Nw7- }ZDDVd O[ ,:'^BަVNR.u7<{Hyo [-6khbM:ͨ*4x=WR)wEO wz>|-濆pC0Cڠymiq"MI HPKlv薖ѕʭKiR#V^ڳ?EQd PQ:Pz[o.JV"S9;MWLB:oJR!Ą*XMh”M SNNVuWZ ,SJO7߾9l"ԯWJCK(-`PޖD>) $YvE>sՅMOOWC4ڳtm.ZT)UZl3Y=HB]_Ow<{Z!KtLA[c7 }IKJ<F+U%X z}0ks5i3+12yVtYfö X{6P87Q6Kԙ)Y ^XPDlWѰXB1^ԇh8~]RGUs(u9mقCFo5qlG+iQ=}~7<׏=3T;Fh5ML,|޳;r_ϭ"Zh)bBZ+4PG0SeXl+n5ETmKxkMBYApj-ҕUHJuɗV-kqw ǵ@cٔtqWŷ6zN1-W򍘥TB.:+A:ҔЄԸ0aKw"Ob|f`j -5DY*& |jyO(DB?#ԣ!=,˘=pBV:I$6b \{'r¢mJE2Ym Gm!Lq"gZ S ^d! O"*f ^(Mc8ڬ;{ʭ*dBON`˫_TĆ ddt D*&2޵q$20PH]]dzY!T$BɈ~Ç^,K#h$93ǯ끨X<G,eP,[D"C8AdD.k- *[&9d'N1&PK%"FBq}6e ӏeMbIVD6g`9!:K#TPiea6U.\|z2n/i6; >Jj5^G!la𩰒QԦ&`0AIi$a zU&xus~zBhor'(HO~~y/wf3[9yMտO.%F+Ecm4y+i.gM156Y9AFcЀ5//AZt!(36ٺzt-8ҍ;ޚr9+rZ$dQ0U`Jh?l1j1IEM̟jŪ@|óiMx%f!E%ji-uhe,A81zg,sD&flTrU=0nӵ\~yNA`z6V3 bUH}:+ )VȆDJG~}ʑX O@ܒÐ9;S`PM)S@L P23]"HXEƉ-#HwqwJ"RT1cUEW(d Q AXrFΩ]ֹ!7m(l7㰄?~Y_8?h`fN65gV}EOr'eIU{ţzQebe>L_g^\w6,t^,USBkfx(uro.86[FHܝkwi:aqwzbx7};?ȤJ̞Q( 'h.G馏 E+i>y\Ɣrjj(l&ԉ*ʾZ|߯E/_O.jHݍ\}Ճ)'?'Úu:]sW7˫{vnfgt'=;=x}w`T*$LhN i\aBI Iꑎ #a.[fÔGNF U|<~\,|c&Z;7xuyVNkPl^pH|z)KTOB3E^boaftC&ǝx|/߽g ߾iɺ)8u)t {L|y |r\Bkƽ>.PzgRv Իdz߿ WQ7[u&[Ń &-l~Yu#od' .%h#]e'釭$ك Cι HͧHJd냳R W5d#d,[; v&:6<Ά߀Zx!i!IBK2lG L6'L LN/:|z6ٔКhbBkP;_FC j4;S;+;^ɐF1ZPN!Gk]RɫV`Cv%$Ĩlxsܸf VgYaP#܍ ?#p`TQ87Z{ۭnӺ-NѠX:dU]VV+Rr+9S륰omf*d7.*묤RmxҘb=e1LR\`yLS6 u&ӨA#l9E% i({fަi&^efqS˗/kp~MD8K3wةmYɭ얭D+ N:#TB)ńeQdJ v!lڦ߆S/Ƥq."ěgPt")1"`" N|LIN\`qj 9&Ch)lfruN+< 65۵6UwqަsW,C7i#M+ePC(c3/2\RsS4 :xYD-&Jl1ˢJp2dB{唍@j˹qqJ:*B!ё &{& EHRV*i& NfNHz鲵^(,j˹fy36[Er@|zSbQ[b(lrBoE9B2 856Cj&e>YBsA} 9ShB(ߞvF9`1Qd(țHE8,hjp_no)Oux;B:iɊXXa8T \cE[ыب@h \."[Yj!+< Qd '21QBY$ [_f[M7guo$rG/-+݆=<&/AzKNbl$a ڤ'i0Edb0Ro٣?Xv:F!w;9:AϦUFoė%G<Hl 6])y øvD`2̈́AA]w>B&/jsXXsE/d89ڀfʭvڷY҇.9 W+j#aSyV||\^R 6i0X]J,OB9[ '!K,3 30MϘwf.g͋_ >Lge6仄즖o[3=L&Zϯ/.yZ\XW08=l^xwGSKiǥ0EN*ܺbV:6/ns>d^-1#ԕFӽa6֤ߧ>*gTAeh5(hI$ʈ(1٦=D~jؖ}7eٝVZi.QrN&E]TFfg+FmTԬk lL PQL RHQyI` ʴb(E9 XJ M&$Kޘ-%5OwH ~m ,ug[S&SΔSH0*Ρ"jo jÒBh JV"lZ7+'.%RAEWS;YdYKL`tL6ŬVD .M{$8}$-PrzŋQ02% " y1dze}bQR  񜄞bo12$C&~5l'!&*`\Jcf/c9ocUB4s%ۨ:M6~j`0U)4mlv>֚iѲ,}56S'~Lk[UB]ewݤq3mtsRCSoaN2q8G(r{yTc3n2>a8ڃ~r6l:o(v6 zmŐ5~@9 6`&xEF*k|H-K(̲E%qmZm١01 #BU2+|HDP9QaRC)ǜjSQgc(%[ k"|I u N6XS)JY=e ֥\ˏmȎ`4?@ncVl_vK>]޽@ѨjQ;EGI%xA ǘc+\+!P- PsדglTe%l%#\b`cq&y leYz 6O#ixO>Itz+3_teݢl4+ǎ:IKo1^k5뢝>БʇGGi%iK )#<%7mFj d+ T><$2 ̣A"KY(CsQ \1ᢋRRf֣fSKLU)I }2eNNo I݀އ'V[DǣF|0`:_3 f`0jKۃtdz@CXvHִVGۀE Q`ۢ9j wJWBpb:$>uJjN/gZq_R#ٞ9Lv K%uUɏ`J*IJe*ѭWf%} Vz%v<Πnqt0ץ'>uE  d9LmNŦkUK%)EB/MӬ-ܔӼd͓Nig 4{Zk{J'{OC}|Umsvq~%ۃdk g޿^{mR\cۣPoά2և{-zpxˋx*wv5u~zONra9K=wp|t2Akz0 ɖ0`B7aT`bG LĄQiYE4IsٴNi5Q9eBL%-9ݵ_W^=Yߝ m7{RMOfOgܫf^X,ӷkh[@(FGEU KiF]\Йfsб<#l)oo1NZP#[/ؒ/2\c12joE1ZT*ĹۂoY-ts컵>-٥[8H4o|"fiP} e1?.A|#䪂K^+^v_M;'\#v%q{tVWk>;29Qe=ޝ7e?>QKt[g-1maC,NhRy{fnWXڈNa 9 4!j>rLI954 6dM@L"\۶PEm1J:iN^ev)[#Ry^؇. #;Jȶ0J&+<%zDb,{'vK djs1 m .Bn&nIBs J!bLP6[QewǓeN`MrE)BCՉLr# ?^1&h'졨Vj ><[{] vS@7:@t{&o.9CsO 69Y#զtw8.8nX/O{RN^j *x7At9~ڥmAFq}ѓ_kL.V61A"kcvX(ˬ^Fթ*Wm(D*dTkl^xpޚQTlXSi,{t0e>w790\gmwdlerc?;[q!fTڙbe *|"ludj-M,3s-G񞣷gnƅ-[r: Uk(ӹNiS>!cox=|ds U5Xk0!!ա9YTs0g"m>x_Sy"߷ʟ; *_l'DHjwD&Q}1Rq5Pk C4"DͩFi_̍Xlª]@Y~bR jbtu`@UF@.!G\LRhBJJA"s l5rsR*c3sKBĹGy`񡻔l*2" +y_CYзy*DۤE{Bud*EV85bUa1T[s49ԐW.hhvAPͲggW*ղ-LrIhsdL ds}%[trnW;CUт&$EYGZF@$ )J1GbCl&κsO;g`fMQlwQ[nIMa 94@BUV"ơȓ 4wE=?EMdJޞ ( 'ld)dAZ!*NP*]R3o@vι8r$ge?S1e2aDo ]")'qx}z+;MĮ4 |7Sˤ]<&2aCbEX*mr6smȊ<XfLjv%ɋޙ>TƧU_Tby$OhU~BW=R -:m4y*ҙJU0M<6`lzvӐZl=]r02\T}׉DK?HHM(īqJɼMrPIPF:?nV| <%|YEtm7Sʶ^֥ ,`p6%[o)d_֕su¬ nTkroeOe@ !>%Z3X+W3A3Ԕ3!hk*Ua0ɇw픔XZOm9iGZDmK19_ bbo.ܽor'R{<7hcj\lQ+L'Rm1kg !U{oȩKUI&!.~R'Q+e) ]ؚ [hC̈́G6;9Zb ,&CkHʇ2b $cokP @@AhlY2߻KN۷-V㼁?^kkh:z3nlzOtdjvhZ, g%W^OauE>f>.Q*HxL<&rfPkL3ЁQF+m;Q{o XmՊǽZݪ%y8# &oF\(d^^;o^7=źuxy:=Fbl[8;ѕ?%yz/zb}tOx}+9IM7f RM%ٵ;:?{vvz ɯcgG~p>8˫aO?ptdHD?݈/pySjH#ml&a/bċGlONjÏzu9hĈj^Qܫk@fK6K{Y==;>ۍk=Atg/@/|~l_^/C~/_2AKMhZ+lyN26 {s\y[|#:f?~aAZ:\,mrV3sTmBЧV1qCdJzI:pvpljaWS_ @r*\@N8ƐኾT #0y &TuTk7x&Yy\.o%ԝI ݕ-+0ye89AΕbc[M>;8В9@ђ =J’ K"ߵ=GJXz, G.f_|Ic< rjuT6<6d+TX_Wc&jYgB&#PTD*uFĹ';Zfj Dn]Vx'EgQh\ޕȭ'2/l+@h{ץK3?rxv_hкBzuմ=s<Us3mͺCвjunM{=?yWiC-l2ȓwydḰaug[lV<"T<~콯0rjozn'cu9d0w'V2q)@TS ^fjhc'thݷCC7[|tk?h\'|rcu4Ѹ2 z/{ֆQk]~غ<}^92B(R1ɢr4vǪ%njGiQ ֮V v^6jE6hBU[V'w͢o7ѽC<` i6C_h{7:wkAսH-GxNe!?틝;B>v"ۨ% WRyj)` s2Y&dwѯ]|r2*g)&1:h959l/2XIhem奷a|dG הɧ ٌMFJ'*arS@K_%~tou3zZŏVSߦy֔G3ClbQE}9 GѪ$f"-pkS;y>Fgm鵚R>F~{\8],miڼA: KB$HI90H?9rƓwz{΃&ٛȽ!$KNQ4BcvxR _bB’ĕ O( 6JA[䜽 %Z9.{&O~ y,9|- ,Zw1 W>X'V{:ɴtx7ihR.h6XtH=c|p&9*;8^_s:G s<p_SXYM՝4b0d1 b"Lqy  Wcvփo5_旟((wC{E/o ] onFq ^eK+!<hBCQMY2r2liĔmKwZC&q uf?~7g=?2-n9:wM])RMr0Rrם^ZZ]GZOy{oI{g$/:L6`v:zl΃sۙX݌GwrOSL^<2 1;=#]2t6zۆLR{.}ycv:C6Ughw? چy&_w;_vfvB!оC_P2)|<q9!)r}z+M~G?^`("c 5K%J%Lo#g%GIccDE`'׋+ꨪ{U{}j/$umO:89$#_VUSEry6Ur|9(6 %'2JR'tJА#?m 94(L&ZgŇ)8G$$6"L3%LT^!ItD}2@M*Sϸ(J- HARNY=C%YoU>ܩڸq{OK լɤm oo/.)enL Zo^Hx#Ub{3@[ vW&ܒ.(Hs( > }?]EZoJLZrлCn'Hq)#}q{.:'H)cAg2PVr+Dp}R𠢷߯I>qv>rz|"{iړ ӧ vtaԙVu6dLycË&:nF?_<ó5M^<{+߹7S1¥ag"Xm\E j~3K+{́ǣv ۝61"NJfeF[|`ٰϾj+.Ԑ.AOu!/m۸mcp/$%ǽ4S/,5B҇TJ]o.-mZ+C014.;)=~4zoP"6N~>'7!*a#czFO n殪b2DZq I߆y@jp&e> nrO6t{΃vqn35ꟷPnu[Q|h XqRyݭ¸Ǫ}wTQ\@>1!8Pqې꟏<5|`1{ٍ|v7udqcߙ69&L6qΤ n&Nkd?/v^j50#TQ^j1[Ň2r\_x)g)C+%hY $S=LZ!Z$Xݭ0c3cAY ="@9xQrh f%zbǐi`hn=#Dh]AmȬ."aQ3 @w$he Ŝf.K ψI1.Pu5"-bO}fi1r!21ڸ@Y#AQ em=CKZ2 TԆAhvI)$ TP,RRR9`JM2d䲥1BuȾ,7ԇR1Ϊd)8$+32`1H>wXµL['DieD.4 <6sTLLZM*ـhDI,*{ Q;CP0JD;˸Q4uHsk(;%.^gL+.QYD82za-4La/=heyF e4e a[4 (& :-V4'spuI5oqɄ%\HT8QW* 1Ӑct`mV&ܪ່ ֐WmPV@ڬQ #GH*A/dW9BRƭX"5R<` cޓRX2Yc.E (tw BX]Ģ_(Ektfh52d9 T@c$ "2f7!NmAKAmBYACA  L}"=&÷21G@lcBgbƤX "zl\R4 f@57]&a?2N@pSN{+ T C'@Ae ^Cv*@=1!3{A;B^ !&K(!N &$t8xR` 3)J;x Xd,2q~8+)$"7%!@0qPseظNbՀufP\ 0?|[LPPSje])S51KrNJy^]fP4\̛=,d%婚%^89i"J0\($r,x$QvM`e&i B(1٤RL㺁 LA˚,z%Ė\2B{WON]iN ?-&sJ uYI J ;[Xߔvln%g9TX] Xѵ]l+>b,c#]r׻Kb:EjC*f7PK}Y]P>B(k|{u(k7Q!۽FX|(#^ȁV!ю2ZuIgB.dsn,jIzМQ'kFV(-5*h!n{mRyd*EPAghHPA~)#Qp;H-2jFXT 9ye|P^8,*VBH>K45kN56C:b8KiY -U+mWU[f~/zTvp^),RK-w RYDf\M:[P0sɺ6xҘ[Dݮ %:(`r ( _cy]ZF Ay3!z cCGeWkqR^8η܂:) k3S6$7N [6;NP;̢ ?FD fh mCt?QAvν I {KwZU$НB=1 t7q}Znj4$ݤ{F;@H[?>yܳH ""=@dgr3i:MOF VuTMμYy 1H]s q5}p_Gw34U _(@%l?GGT:UUJсE2FlЭ,GHRVH*GvDڼG֧0[w4he/Wͻn6P9blKMoJZ`|ls ^o+Ԯ`탟x}Okd:"ݙHFterwD qͧ~Oow6jk-5@[d3.88co[[ f4|<Oʉ߭ +AƉ{gbPTmb֕Z#d޴0izT&~R4䒗͋+Ggw%Q~kyO?VsD \punu~&/+BM*덲Ghyhqx[Ɵ4/o~xq1?^~r;9BfOÇ񤞌/FVnM@&?]]^zhO[؎+fy%ˈݗYxeU0ɴQ6-O''W=\s|:_ss0WnrUwܫj4檳.BGOYWnC~G.bV_sRdQC_~Ջ7޼x7/W?@ F`^.|jKzKKnsi7]/rĐ㺟[㴎IzŇ,MǗӳl,j M.(LrBo 3?],gTH[Pyg*zQ+$%46$Y*W[D5:cxYmW#rW3_֜"rgWei?ڤ]om^6SAeg!6v'a(*ޗXU0?22 !FZ&SIz3 *4_/-ɥ*ɶ+S.n;,$C{Qi;LĠ*sj+).J}RAuo)mxVkm_o%e9"~*!"`8-F@֜ c=Ӊ/sw ߪR#GB{F^;z-MmoZ9e8#"փgb|=o{i[t:{ ]' 3&bAqΉIy)]By gdKӳ{cCi}lkaO>C!_<'eՐ@&OlbڗA?P8:[BSC"pj/r"Fi-6[l ؽ;:FǣI\,hߟ4읔׻\Mw -[{ n]JR8&k=ϙٛZ3^y41IHfFdoAO{Y%3QBvRU !RLr [Pa""H,ecUBQ7Gg< \)59C`P:AHUa1eDi-g  B<ɖRB<3p]iUl/W]'ST(Ar%K&`U,֘hUki f>B\ubcZ{O+VcdO1ѧl I֦ sG&듥ExEښez>i_OJ#oߍj'K;yh^v0 F3x IHӣ'E1s79 s4~^rdlj3c?'s?39tqfb(LV:Jp#? jtT^V Y՗+R~Gasi쿿/(Nx-DdO(gj״VܩOgUxmGO~ILM) s*k%ҊO[O>q RR&PkQS??QVzyMz+{&낢 *]yaC9rhVGr,MaÒ]#__BE:;a&bWuҙީҌ4#֨9w-WK]ӆ)/e]7 aKKˇl/L덖TS;;~<zh4a܀!_0YiX Ъ %iOy[&f`]eC-=X;[jَ6Wn Le@+Kgaյv7T̳ ;:a?[kì"=+ nyPaP?li͜܉6O}AqtP~:|#O"? sGE$0@+8b#bG0Q紲QQ&Uk@ZXH$.':Lk\9і8#)+8Rc_cR1n~9ixE[*ޮӇlCBm Ǜ ejuutxRpr:P԰)-&e1f~5*jokB=P~I3Lj8@(Ӣ5x P8X} ~IjAR1#5GjK 9#"$)k%q4/0f,=X*ٗ*lQm#:%o@9B9._'H>;RkZ=P9k"QE$c8# `T",fp2k쩳ո+k2Oo'&[["vNjfּ ƍ+{4+(w(̸^fw4]7=67f5~:k}،#w[w^Y}yI㝐kv)Kl^;vmmg5-/NxK{Mn_]4`{!w鎆kjKޒdAGlsN[uǴ`[ձu[/s:+%}k{.ɨndߧ]SiRpIv4a/>/Y^mym) Z29HlfJS8.Š0ζl{L`SIolV@p 1j0k.@OIvki!# ԻhX:ʥg;BDF+n9*% w6@N6Sԋl.Ü8;;/Px`P _9{ 4Z6fynGIs篊d<o P__EaQ|SDb7_֛zśorO޼)|hŏu`(4v96Gՠ> Ovnޭ +>"}%CɇXMYq|V|2 B' ϊq717FcMun3Co ?S#PΦ0t5,J/jP \^roT~ﭛ4]տSx!ʨDM<M-2jѢ C[(ڛ 2Ą21MpL$TTl/u84.k>0(Aj@ZsX9w::J,@fZjwS@3yxSbx X+VW^\Ҩ?ͩNzM@g7Dg"ǥmPN2LLw S5r Qړ5L&9SV_rPr# ӐW̊yO kr`n_6KhY0xU˪JُXPds@TKɞl}nOI'I٢8>.@F˱RZn!PibJ(OYX J$ ^bI }t MR$Q +(gNϝV;H]XA v9Cs4gN/SLrIkET`NI3%eJh4Ҝ$""4\)h×(nFnpj6>b+vd=!*:tj˵Np Z%k1ZΈ t\|4_fV/ǜ+y ʪ$a%@iR$Y&ŌL4uzp]HLkB<.hMP7$IE̤%1zE5GJ$c–ꄱs Ta\Ø2&gƀ.U&$S#FkuN,D1A"8HqH[^ ?(@A? _;BGSX'fUؼe;IBC PDI,,ȱH' G ޑ.Wݹe[ :P't82`g4\p[0mYԡcVFoM尿엫<_/ZEXA+|7/P#`{dY61'2 #ƘC e!kSs)7cx-!-~n?nE޹M~٦b!BZO49AY[Xu Uy0,\ȳ62ȵe㲜 'A P9u*u*>{ΐ>e#hI %pDL!ODO,hn0lYBZ)^țפ.˛tJzږ+|^I/ZZo39;߻_ mNOPQ!üDֱ)ly:7 H݀D-³'is.D1cc9C{gf@PI ^K"X- ϒ-:V12b<"gwfy=ggcMKȽql/lΎ)_& TWW?F|>`is }&Y|fuͭ[Tf*2SDQ%Ƃ(Z]=sՓ4 boQ#_~L "BXR˱7JRHoHU}(sg<+~@cjyiCgޯ)VN늓BRrutXV&].Z\Xo[tjPgZ,(gAM|:IP:2A FPhxA8\R8f[kY5v:*jIq)54*]tkwע h)T|%J)n:g \`::̜僐}4tGFx-/ί;0Qyź}6md4h;F~T|˵2(s#v($池8y0j O c)DMTerʬ@E"+ֈZ3966jsr|+ @y9# fH](fSXS^W-ك8.%ULƔ!+fxcAaq("`D"BmT*Hˆo(՜))G@aSʳ6 &d [0L* -CIk&2G?̜pu..f:aQr.J p.>lQ(E b1&`![%dJٵ.>. CF'IZ|X\wu5YΏA.X=!ӏ_oLI7f0sD*o4.3s51Lr*"GM?tפ{fnl~RuϔSbƱi+*JḂTr*1E'5I-lK֢RX)J SXYӷuK}7[ &y ͳ\y%y<T~{;^GަַRBuH TSZZgC}LctP(Z}29B K@x5㢔"E$%5@0sMg,ǔ,4_zd@D-ƀz^"UA{eb)Y$Y U]#0U'q?,hMY)nQ( El}߀Te_ɠ=E(Q gܯ.4~K%Ph0r ty籯mZ9-̍\ ;.̅V|]_qFً/?77kH纚eųˋ؞L/>3U1.Vx\7CyWMo UPwXk:_Qij.;|o_]_e鵩7Ȱ,o5u]Bu.yree=[w~Yٺ'ͻg)\_,V_p?HǾ^x֏?õwv٫g}@lLQg/>r?~N'G<γN}ZN?}]~!ON~w)wZgM};γS@Ik(wzkN.xnd\h~bC(\C᳁Ox~::ŵW 0y.d?Z:)Q%g5;ZgUG υSq9.)K+b}6{ L#$)龲:稚r7V&+en-W@nK^ܼrɋ;v,n&_io^ ko`Ao\$G2ݎVyNj츫^:7y6zϊR;sTp#NR_տ6wg|9/z‹}|/+o<G!xrk) KᖂykoqVKKz}%̒ cG[9 r;׫/Y| /2o k}駻]i5ɛiz&W?q}pL[ɉ۔e_uoqm1h6!N6[ L3vt&"m=>wǍAF7<-/緷ұ7{gs>vMVYO{xo;|t>❹SMX>IUV}{d?,vIhhOݷl TթK(dX @ʇ'}LqᰨerϹ|r3o =w+rm>|r 5j+Mr4F;iȧ`Qg){׼D򙢱xK77Ok4ރޭ弖7o }iEޘOnR/ViaeBtFiHrƒ ZY*.~)\$B/$&L!lEA;)ϪdtB6s)kS|A2-ۻcS8d9ZLFZi툳"S 97w/@b5rIxnN9 ə&**Xj! 'ZK!PC2ÍMkOfplκGlZN1*υ}&lhB0`]Đ0ڄ{B{kl(cA4PwN (d.Cb{+r&ք[e1{Un5 Ge`쁠.WocY K̤3i˞`<%CKJ6d_ fl9V9iĪf.19ΩR#x[5D5r-PZQv0a۬# #0r )VcrGYG(s52D8e Nsp#9^cO lαBjՔcF< I!sc4@ĉ}=b4V`,)j%D:*D{ ޸TTn LUd˄wh8Z3e8R4rBmQ x݊ *+p-䩀sRN3qexh6'_T Ly2q,jӔH%e#Zkr46fGy\bil<`\1TFb*aAl`!* \DWZ # ӟʦ"(UQfO9ez%둛@cngdX2)f,750!Ю`\{> %5,C@zaH#5 ",a!4KuYL u^$7[S 49 C ӄQ*?%d[-zjbpVl@մ VdB#"}+޶HLt&%Ji7i,T BЙpƸ. Pc*A*)dgC( ')C@B=Z*k(vV(:h̨>CB+iR+ A txvaKy_\Zv3 q }ŬـdblC~.a0! @߇ s#vv3l3xa߾_\朊~y7_2δ]@B6MP ckPWZq@u,}l9D*.G_D@˥{ wS yģy61ėj,$tB\xo!jITx"2D&jZ:# XI֊.Lk>}BH$K: Ec>(2xpGx!cx(`ZU,*Ωj5jlj)icu1FǞa坠NbozIZ=NȾ ^{@#DEDF$)1aЋ:"Jq Z!IlFLs @>`,CkZ%FLXI73j!h[/;H_,#U_c7Wɲ:r`ԥ.d;Eq`0@ҿHYuhLЭ.ꌥ,0eeRbrAs=?>҂ Á x0&[N p(:# K%њj0$;0SC@*` ,-1'0H9+a%D}z% +Dldy$:kc%5`F”kZaX aM9*4KX ư $> qT FlY+0 }JJ[@P?0 !(`-I{1sr-l̵Օ7!M9RsUH\ va.%Cz5bap˂G-1Aұ.DS7L<C ci'^BOnGA$lC1AGI2Msh 8\4.Hիκ||u"" >7ML "C }O q18 fi5:@8/Seu@x'Q<P VQ/xrB8KTd>*QCa˭NѨo߮L ΙO Ĺ zL Yf`d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2^.HQksb)|@ φ *@ F&Ћdi#)E&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!2 91@`˞Ȕda\no7߾Yf FI ƣc1 -@&?O)\ƓO"ͷZjxh[2 !MU|/0f_gv뤄+ᬪlqsؙXTp‡ |fէo]>T^yT^hoC~? ӓ_OU2 ̶A_._s|OtAcڀw4poǗ^Ԁ "?񼻋Qz.Jp+1 -{%-"IH,ˌeYTy^.[n14K1_t=4t0eTd ,Jki2 |^c\ۧ8;m!0:o^Ut&M^yd +("S8 TL.?Uk,Tx}+ڦynKuha[ʙmj[mGs'z[&c-^ۻ,sGvĖ[A ;JR*3ghiwzv3l}k'@ye 5)1c}&Gu}: G˨˙٬: "%υ:mt\]fV*ĂQ19.wV-g_hih|=Zw>gg7=+p.|!Ww9%zI=% C׿]S mpZ*WBmTt[a+c=OFӯ=o'W8[$Zn5j"e.gbZ*OiSpWf |n=vj΃}؟+^4a\YL[cӎlYS z*ďO*{V秔Om5DZߴZǒٝLo){H,`,ASoZ+MEW a5>=9ɉq4F y[x)s9J4ɛ-8E)ʸ4ޫm7Yz";B6\}K7.Y?z5E7~FqUG9Y;IzMUNVL^h¸/~׾.o3SϕB*LȀ;!Jⓠ_fsHRyѫQGMH=Uh,OX:R0$Ii(&@)$f"bA$հ  g%'ZSe/SxOsjw<3? P=_`p)XŨ21$qAr IV)r@;5hG 膳 Bh(dv֫˛,tV#7)HT98NԶMjMWϥjno֢v t5mpW/V{>h xcYR;`iN%8fK"CfTSOdDd|?sg^z_ U/.Q/u~ > >gwq-A68 Uܘ)כEt}4:K+{BYOk\R̋TɵgmCߌчh (˭ g fkmkk0Qֵ@T׮NO}kB aނ-r?…z:N^\-j5DmFk$ձ BGT R'{R (>24HW3W)J!Pjeo}>tz-|ܕ`.'NrlfUd>p)3..#s ì%Lm}1c7qv'NJ&doJxz3\ Ui\7,so|+}"ʾE` qs6rs9C4%)Y&b-zlwN !'~, z v97PHg]|M92oTS7m0YYx^I囧 zUN+s0t/ uVgisZ.A$Ir2\9DL4qo l:g*E\BsDֺq_̚'3a;wVgN4?O,Ԯn[vΝBY]6rc(K׻kx:z~FJPrF_Hy[孵߷F")YTN,*ylpB`ҋSɟ2cOIW=~/Yߐb3&"ƱB%F_%R!,.Cm(vgݛ8Ohmӽֶ\N׭ۺ7Ѝ]XbۧϹfgvM(u"-YdŨPGTa)TCΨՂ2{URu%.IVg>QM`&'G.FqFq=[po8 rBF^\{'K\ +005#˶eWbи "673g}H3Q¦J(+,!IIK ()yb%Hd=e?ڲ,pZd53 ~Qf-#RǘlfFV2~Nɥ |5Zj|VPcp[*D,δ`f+2 ^i\ӛ8;] ~_T4]1tr=Yd6:E#ah;]/xZ ǐmQC!mR[>[m5:#z?k9 XD='QJ=j ER2[>8W5цMFǮ3VZ{%άVE*P6p,q48+N)u"DQ[q% &2FĐMq# *UBJ+`|. t}ih:: mS(XEM&nw #yhAa3v1Ť-׆Dz _>Rve[[\@MJ#y@z*9 Qj/Av12VݚHGI̷!ae2a2h%R. `(c EcJQ=c2֪l(A|7!e5\>e}]ZR)9e.B7(DGkPHIpwƙ*W Q\ EKb1֍s`I4m؋qkٯ{9J$ U9Gc4I!84$qYՍG1NAA\ֱ*N+H\)m-u 'Bk4a@\9jwnrNJpOaNUx8l>uPQp4HndS%Wʒ@I^J] ۀ&)nh9X 5rSxCોכd 5׻|=l^솃 mżj̍[;+AHLh3$a];&A60W2 w0b>ǣjgsC&ݳY7kZEu=a!ybx2JTdO1=x zGeZ&~[0!9Oo߾o7_}G Sosj 2 ~3>|j+[M͍8dj䪗w^Jkr5a7(2ha}V IſoKD][|r}fUI&$lBɢ'7qS8YHvvS;V"'+fjy+m'嘊0Tq !Ev$F!^3O uiUkc4񅄾>vӹEBw't+ubƗrC_WtAs@8"7dqpRY`R; ([=Fj d5S^%C5MZˍ@y^lqׁm\ ¾Q5%S/GQOS:2(Zr'VQ8꜐1~j*n-/M5ΩBBC"F}Yڇh ə!x=b,9A]` 3qcgֵt,/3wliPG=v 9mշ`Y Al5vMS[:lWɩ"]3L*#Jƌ/O^@IiGZV˭Lcb>-S,UNI!=32):gre,>4AnzT8Z/ɠl(߿!Ӗ`fyy<\_k$w6P]ؔݠ3fvh!eo+CK]?.w]_inVqd#W=k{@xCycnz;୞<^!4PptpH(Zď˺A6k\-5`}ʛczLޟ-Ieu3PI^zG?)kYjT(5*0–zy^Zh4,jY>jmΊYq1,%7wCK`2fF|ǞC9A ocԚl&U`cJdV` 0h [%>Quwqg^JmrQ(V;d[kM|u.`/Cfg:E8xE'5m ~W!>#ِԄ Q"KWpY1#Pr$O6nt.1 %<L[ۉh=(A,!{WU@{B.:2YoF9XUYKo_L+Ɠ*ݤw&FE ?LzT#&d,F08=iOxq) a^qd1{?T@pq1x>a {aϐBF0c|v9@#F@HO;xR:0|+Nj_."K {!+B_\ bUWctRdJ)5!v8D|0\L<< jͣ~^Lۑj_?{7̛QԾmkqo!.yUo.;ڣ?mԥKN*Ub "$6$'߸[3Iҹ4%Z%'ߗUޢ؂.ˋ/ݰn!s$pRdRVQcuuS*@KfhP[|^msup)]]ގUIbQpňru")(ݝIQ۶r}׈ixwy˘d،\pϨ,2YmUƻC[̜Zwvxn0a>rYrRHYz:_EOٯ!aOk)sP<ȭoby1ܪv%EGqn+i4:JT(U)wQ !e`{"TG*MZm[WΛ7iE4IU50fO`b#@&B5sFgI|?|0d wYp[-9,_%%k$ʖ`vw&G#jf*6!N o3@#@hU B9/1xSV20-֜M|H~J*>νR9}>TIVuK/ ,?ߟ_.mP j׃EӀ"էێM:<͈-AHr 5s㉮-W;KI2)5Iy=U^$es;yZo0;A(ZfW+kP}Ҧwn;d1\}GޏErB9StB68}pBgSkPq+h"Kܫ;JӖ䍋K9S%s)Oڎ¼a9.?OA/Roپ=Y9X޲-gVML`[Gm 'V88 gZFB&(/HkIo/ M=EuيKl7@S7+g!"ҧ2@rSz9rLk )9+%pt XD'Ll哅-5{6`fj4n*ݑW"]Hy?`ɶxaʦPE}(jLHb~uwwNX0()%|^qT:L{pF!Š>"8h(o]P ,b4H;haW9̀G408Iɯ@F=PƔrL,bLʍ7lů\;[1O UsV`u?7n0nOj|=T%\om?V}* eǃxԤjjG-=}?rK9 = .Q}FÀ"6ގ<{pՃɯރkmIiǏ-6NR<;FjV]!;M^ F?V饪V(/ !~N8WÄ#2"!lax?h[>f͸Wpq0g#|t{Y5_9mbJo7S?zOuY7 aҾO86!|)mmO>ճ'm2LӦXqƹG  ; eB{פw&*&{nzfVtU RpC'FBER`x䑊EL@`(|2*X H"{DW5gDs?'&+' 4/eIåF* vY Xc^ :z}g\;řˁ9JN' Ɠy;>| @u8Q#v$y) v)⼾ 3[Cs^ߪ伾U[ʣ@L` t$R8*HTg4B ŽH6 !GС"(-t`)93*[(6Ko(Υbr|t$ĴBL! Aۛe eX l4TAUg nr7Ѿ~-8)N uV!* }SA8 ޶tl;{X[2lo_&9tRxv|2۳MyWMkګus9F܎T?s5߸5~ݓ@wهbVpWax;5z46%.(~?xQНROJQ E#o6F&ɠFB…:INQ";L|eHGn `I|i 8t,g,;Դ"XĔ&RsΡΪں\1ʥz7+^_LR*sxQߌ/NO'U6^jd$VKQ{c2 c  _K,Nckt G/6!f>ĦGMЧɯZ\1?IUFԌTAH4i缣ƒI)~"=|N%}:(j}%r?MNToyylCh& [ BX,UNI!0Τ]ӜH H|4Q".3:?;uzq)wE؉}?]MdƟ?~8٬uĻg+n?qxzcd=$Vcw78ku!߾ԱpnoP`#&)xPL-CՊ(Z=u=^KSh$LR2cҚKH܀p> Oڮ*khGZӵ`嶢8?_]櫗2Ցe@b*gZV6p˓,%iu-7G:-Mw>}IR~6k-'mf[MЯV ~st1qp V!;r *umA}j^/P{Ze{*ڄ ()O墜-nN̉ehl82D!Șh`u:%*EKZy *wVgjX='-|f?qy qyV43_ v<`+A=/Vztzp CNOHk-vFB%P 4OqIs.10!eb1m\9vWt/LI$ףݴ_.߳9Zzh19o Ͷ\_r'Uݜ˹UxM u"xrF^( A8uXʨIop93ebY*s1.gU*ku g)$IQQܡR%O\X?ށÛ*Yڛկ.ykxE,_YjCB B.k?Oܜȼ %li5%D!!D)$3JyBPbPnٖ2i|j-0l:S߰Wc=x;ز?!R)lH!m y,$Z͉(=^* 5 Ye2(DVG$qɍT 51tRy^Q٨I4j\Sz5{~_[ι,l 9//'ka5=\LI6K÷>+Ms;ˏ^{Iyl8nE{EJ.٢V5#o3asHNKo#3%:x9e*S d_*eLH"oA%TAd+AQś4.*m5yX$c[[ B½Yfd"JmrYN{a짻&Ow䆺M߼sMb4 z$Zg}\x3%Ⱦbل'ݐɞ pY6\i(4CۡHh?rJc"^ _K5y AbIǶV[ں][/Y `y$܃#T 31$Iu> S(I#@ȐfyhkB$^ȌWcA<]=by6φP,b1"t$^shB%R1f4ʷJ8âzd`fZDTeI0~ 5HaӃF*-Ad*Y%}Ɛ*txI@qϜ^o1ٝfϚKsjKTC` e$@ɼ hK J{,ɁAJҩR\L%! m庝 r](99Mgxŕ4LlJ'dQ=Y&r2y'N: 6%*m;rfkLR&%(uAG14КIBZ%UE&xU(^tF rt׿[~{6Cmrŕ^ |ctTo)ԦsZQç7i SiLMcz_TI] T/kۂTOj_VZxM ?ʔn Vg|ju p(\*4$\@[)<{9Udw<\ݡ/YD "rpLC`&%h`4+!{A(VE볈v-芋:֙DILTͩk Mrz%HZ29 *+)@:0^KNvZvH|YWzZR,P`󴑇ˡn͚eo8G5دwm!VvKofuەB*p!륋2Te&dhlM#q_t#b=s.D$ $KC%!Ek͌Ɍ3Z!*oZZ๶V-EL0Q[cIdXSd2G$%7ldhD+Q_fٴCPT6b݋aUe.JIQ:KπxN+&A,"Sȳ)ŤV^.=o{>;)cECE/f"я&*EL<$B zԶU>,{a1 BF$6eBq8&t<++X hZΖ|V?Մ:>D+(k3|d&QOiaTW:'ĦrH ,& |z+[LX*&$ţqBAfM2l /Ow:̬ʈjK{;e[/pR9ILG'y2,ҟ. 0IF1o1UWF]۲~oe\p;R!Z#Ym*8πQ`:{dIcޗ٢Gv=ӋL-H][ކ=Jw[2J]VicI8$"XEShg*w:En4rcQM9h1%I5dG'F@-T#W:1<OqTfDϪ0nC1d#B%&QX Z;eJ)s4Ez}.*¨z^E/@ V𥜡 9$xBJh`Pxfז.#="23hIoTdN_(fU1.80pMO0yObzJ+'`0_8BNЩ&˼hxWgD??ߝ?8y'Hv`j;OOM0/?o?m M-Vn3msՋ t\`q`k^øҬ5 qE/%D$NJ̱:I p8fbNH&`R)h)Zޟrөh⍄>tm$]t yV;/ qz=^̝Uǃ~F?* k{0Pw23S8>'㎋_i1$ \.vS?ZT* RQ:M 9hF'eLL$C$?ZLuIr8ZKaM&eF?R`.PO^2`]>̷3s?i0@v=?.`j{j >C_Mj24,κDX~;AJɸ*!D(su|UM3U)n$+ݞIv^Ower<9vMi |LYѩ(P!\\TpU檔I-Noy6\N٧*u+e\kw;FXjv.dRdEXgITy>*ƶ_6wyrsXW|t>i!(@Ibh&|eL*Ռ~ӱHΥP,=cHR2pE DHgC3aPITd@k}E=w l@E*y,yCdV{Sa"lm:;8r3>+p{ܵǛO8P.ؗSogOҽuw̼:}/cU`ҭխݫFmauOU>-v0+ oI}a׾//WtoZ9'ZWgPRWWQ*Ʃ=oTt AHdJ)R"eR\Efp!IIu .zڠ%+'Sm 읚ބy!F l_&J^rmZ0Y6i++mdȢ Abуu7Ya#%:i]X$k1gWiŷ *pksgg#'Zw =V D2 1#P2JYJd9ACw*峄֦Y On4LÖ1ӛYKq#0 G%u c3Iቶɉ,]JzsW%+yà^?ݙNxx s7bOxؖbMDS)!TeZBDꠍzC_k}LXyl<0{`-.f'z,Ib;}q*U=(. _^b0)ΰ3MWsK+fr^õѰURf7Eyq2.hCY;Ϸp o3,l7YI-^-]&8v6-mksץsK;ߘUx[QZ/yvtjxᅍYwH-+jY-un7m=6~"&.pEk-7d<5oC$߶z+%_}ㅷ䶨]V4_6nCY~Ȇ6"u_Q |!odq wd964ٓVʐYIK(޺@6hRc5Qߠo%fz.W{ 㜃)!O:)J&`*'Qr%ڔd9DEOtl7B!^x:ORB\K ch0%z-mFWÒp=zkh?I"x,G)E A"jO+MN >p/D4,Dee{V ߒ`mW-{ZvP3*5԰ѧ L~e*Hjmb"KI($t!ؑ-IA+th';\<,QBƁ#q ñvov;`OÐmwyđhj$6 X$gͮ_UGE#dd eXÑq)ں*jG̰S[kS%O''uFiݟ*E:q%W^q3v$hbP:6cLG1PH,K-"hqV%!/2hMG!ۈ,PԵf7ظlNdGx !~zp;U(҃jH"Q{DLA aŃ 8%Fz/Ʒ[yw;mNh aY⫿*d)#-ÕRjoOe_۷4x쟃AVLJ'ͤf&oû0Wq٩N'`?`8c{:,z}잓bab _p#˼e#U0ɴ_yM ֜K$ 0lqN.Q`c@3Fǟ_3x:{vq*NUTTOG9_<5Ci%se|賣J9Tn9cQ:K]R0q@opϞOiYqsZr~~/ç?`:C/F54L6-xR)b}n'&FS r%s;#r'J=. H%F װF0K/MM g;3jE8싲Ǽ(s|fKv[C,Q_YH뺡W]oi"nQ8ܶvDeXI+CodU34ņ4N z߮uʐX7F@#~vyJ,UUꙫ&.K;:!֒K>/N!Wխ.լ̋3RMhR;6;uKO6ǛI>i髢#2l=psbk|zS{υl{&VpWo )t4M`Jm Ad٢1xoWT jSI &>{tkY1'$"Cڇkv8<9yYXnvB1x!d1t 3SQ kC`)@jpػtc_::;<]{(Jn$dHΗ:&7M92h#Sntz] ,qǏgUyl &)?=wutF[p0ŹAq,B}}]IGwHtWXZ "%.  LPǢPH0A:d0Q`'($t~IBN *(6S[}ŧe/=BWc*'˥j=w 0 .:wȜ9cs  8yȌ1ed%}̀"+lKZՄBj5բlafԈ"͌PYAޜVN,c49 ;3@}N(eE_pd?^JsR5qPk.$%լ^L"q~21QV~fܟJPo~#Ou GV r"x8Iٓylm$\hχ#_ ;&!{;=33;dHp 'DO*eqk  jAs9#X/lޗpXCRq~ZGͧU/~#.Yݛb[h%cB5ݙ/'@EzZ S5N~FIGCL=9>4H$VLEa+|l?;|pJِ%,^tY˪7)򬹳ZWkgYvR–;ZrԷ7uJ[7E00򞘈Va5\:BAF *"[苄u+¥&rS0jveԅ %kC١Sm͆ hRKO"C B )AsBH {#}Gش FdbGdTaXKJR.!s0{!n bkb|M~׆G{v{] OHttoT$R6lb4p&=JoGq0M<",_rvFp[P,'DٜHskY1E8F_Cԃ4.RY qX^57:{W?xү3VTQ~~h>Go.&^I<7swdX ǜ5Zr+j&A 2*`)Q1*8ӏ>m?iZ{sْzxcGh,2k!YM__C-+[]J @!pC\`& GRor,2笍85eR[Yt<pkJݖR6ὑ({?* Pz.}*$2jk1#39w-JTxGDJNc2)`!w0prTAP8@2A:P6 'a(;̺w6Jj#TF[95I|ɰ ,ݳdTZ,6^MCLQJDP.9ANAyF5HR#3RÞ ) 4 (ʑ[G2)*Ecw!}ℬ:xIgelby8֦T,=4yOևeƆO=cŷ!Ob]f`l 63 ލ}u7Zbtu="]kJ/ڡ[%Q*x:χꀺ ,rx@Wc+@6i|]oך*Y_]vU 9oqy*1gI$G߻DX46rBP.#(H'g( R[0B(#=q㽞F%F id0Dx@B< p=1%R8&\sٚ4l}mxv-p ~\7q{X2F;wݹR}q2䓑\2̔1&gFӜasmʥFQ}npX/va 氄8,mE=T萊@;F!(1VX 280 ŻͼOW'Suh[& aS>Xڿ.oVkTۘOc<* L e$¼GX޵6r$2X;P~0X$8 DHJke E=rG+֢8_U#mx2 QPRߒ} u5ΗCK`_-uC:xmЗ؎W}Hk7-GlW3uG2\=^$D+( W .'o8$gG;eaz2mC>RxA"`Apa$5(,Qq8*г7J=!:Rs"J↭mWBs(Vr ң/GҖSy0_o,#!| @2ʬd_ѠKRg.h8?O8 D1&џxXHG_tОF'"(_F8n FcRs؈b'}ڞS@}~^-.I .i{mZ橅K"-'8ZQ_y fyJd VEUD%41J<;&,NܙO+lr!NJ!Pr.h(|!>};}" @jhdBSb'i9D-9-~~jP5ěm\U70{$طu?j:C먱rD&jP_ݞfcNSˈA3l|ܮy)]ۜՔGl.ˢ, ]7ES?/n*zj9(קkqFګ5d/dX\ 1w4ZmFYgCO9wُ{,{-:ܧR*ɍ_In]cHg]&mg$ ?VOH_S:  ZƱ 9nFUh~v$.ǔדу6ӔL^<3 ]Uuv6o{ҖͺJ{d7$s@б@ͨj–pLMx0_Sd~ n: S|4Ys'd"C^Qe4Q'"!Rzt.Fq8~K__\ROs=sؕAY"ӨX,ꘃ/ma889.9HD F`G׋yyPBz㍽Ұ'}zjҰ{ą6zeOC;1eR]+rbzv{R\zHŦDJA8F)RzyL~tS#ԟD:HQ!Z¦H$D#2 Q34U9&Wk@Fq4RSJ1.XDn$~ϊqi ֞A˛mcp_v6?7%;v͠>A=sī/W9#TWx=ZWդtwG!ɉJJs̍'A d"S#l}&ݶxCQWY~BExP ,+~xk~ڔyCTW}IkF}X;㮗xssys|q T̸}pCgSs+sn!^g=?Ipv>pz~K0Y6ڣ B&42a?JqŔdJs(3.9KJ@2Ih¢,Ӿ-{ڽTdu$(/SҪ0ZK&f0c" zu.E 1W\ )K$)PN`9X$87N9}dJ*KK/!2eR3R `r7=]4Q .?5SRnRMrq%ܞ7[ϼM| *c.J'ysS .P|!),O–GQ ֲRDAW2iBTwt~O?Gmwf\5E,;oì(ss!ci ܭS3/xWCSJ73NEb(jP-+7ʐni4u k#cBO(y2ȁD6:beS9qBŽ..rncSnФ6F?>J?ޅ0Ynըˢj6`vH9h5oupG;L%%dRhlmɒ.q5\ҭi0HڲCB6k_-YMmL].G*[Nj>JٺnZG3 #]ymf~UՆګSλ<*:ܩnDeg)8WE}J-h~Tj쀚7{Y?՞t;M_uhrc(.`2P%|^qb8.-@8.ŠzuylA'k攷h˵uR(I!Fc#'=z#י(7]|5 I0DŽzGSJi@ Q,]gLʝSnr}W4l`~QkFӘ7śc('R;|^I5m7|Zۂ.om8"T|㴠$#}W7zm3xM͛i]=SVN߮Y)Pv^'˱W0ŏ]I[}x#eXz[RJy(]!ڪju1$FO_.bsh(f?$Ktɼ4e|?gi SOj쵨ZF[5C8AC|ϘE-\:fWoA_j*vM념uV`<=:CRwaAR % Qm3eZjx pEFxB_05i &3 U5E)X}CU+B;;rI%c4Ձ;TTZOWke]PZ  &*#,]")i8([pj k$Q>8OMO9, c~E\ 0@,HhjpVnâro"B>$BsFy@4p'QDiJ&ex%-L&@91KG; m|OV>ӻ.6O^tmke{KiU/Ji:iK)=AF!RiRU?{Vd}Y, hLeED,%;^ߢX֋%EW-Ux] X{c(}#/}ʥ_=F=;xdcɃGpZJ(Lje%IDm/s YJך d+k. NU6(UdW,!b)9Q˲>#j|*LtB[OOf-#9u(4[d;k\f]kwj`vs #phxu]gyXl@NZ:"tى?:ON)ѾsS"z6Jd] ! Er;RℱP.K[Ft&T+.-dQ9(Z ŵ80,8km,2KCFҺVs3q6S(tJDkOWXPan,M>/D}'Bvē}95:.x(2t :/v2j-1@_/p"{ؑ= &t(2`Q2RlLhP#%f|F PQxgMX #twO.*+{Pl@[??} !}S?{\g=(Ys(06ő"xa=` SP<:!i;"3S'Vµd1 =وd6,IQB C5#gEţ}aqv.º{녜.2$B]#Yi֚zr5jC̲C%R.Z+N2O rPjs珹.qZ]:ICZ~E_e#yLQzȃe41gT?x;6O 7uK?_?/*lweԸ-WUf6_ +'ҞR;(}<ЀJ墑YAC˺:ïөR} ˦H__rYl$CT`@ZgI)xW0F׶,bJhr 5 X:`( j%I{S0 1d>h&'i:3wc0nry<#<}cӟmw'Æ)eWΔG]R=Dh+lӶ*o]WͼJj 5z0x{,oHGƌu59"JGBRCs>/)JD;7QEWܟ"o ǝDԮ>ч͏KF=t5(\tp]{d۳WźW'\mt\g9*׆ b^1AV%LmXTVYJVƽ=wp:9렻FzwTbV;|O^)>[ۃCǣI6x셃;`WCMTU ?#<^8 y [ ,RGUk4E\˖uv鎒NJ*/j2a3C6\f=6Cˈ˔G$?NUD<(A1VHz6u&E| !U>te@yYk!-ك1oHy6og8-Q xryy]u_v՗N3 ,_ȳ$#k\܌_OsƏ8."O"]jb#t焠.&|jEP;:81%:rOY;8cH+5[I-BNhRvEFH!L[:jr7VC .mTHч1,"%g T sEdt8[:OBp[> w15ᖇqaK d҉h]Pka,j~Y$L3F lێpY5ߛ_%bFbZ :cѴgYmy҅Ϛ]zԬK3=٫}Vrxmדo\5Z,5Y'ޅ'چT.K^cs)N7YM)=ejuhQ$fI1ZWɚR AX*6̰0G-3&wbaћH" qfXL3Bj£bDul7R_mnf;6&O.W"@)P0$v;Jc=jQ_01 !{_my 4Tg')TR L$ H¾Fa\D4uVlFp9N+.橠v3Mc6=j )޹e58!TD @JE*b|th2>CffQ*ƚE$1WWRj Nu63flÙSx*L?EDm=">&A%#'KeR$5T J`x ]ӤKde۪lxBAF dLH=L(ȞD9Dl٢K¸N鬳셋qS*Fc7~B@`gkb1v:GW5j%!1pTv싇oxxmzwMMlxx9 rF1zG?P Lcě[wQ?9EԱw.::C+bx'w6K~r%K 2iK Bm&33A yK@=ʶ%9ġve*碔vP>DjMe-Nr$2A[KY\ߢƋ_&AJM Hk@8p!#d)/f#ȇqxo3TT;~ Q(mJ6Yu4):8lq唌ņ(WAb Hn5@bc`9XkI*(iQrcaW>%_O.G'6ӎRXi()S orZj-kaړ,5U.xW5A<m >S3\˃9zHus=8%gԏSzяkƝ~Ir[V:=d`[d_/w'8];lh޽!3[Qk&`OGӂO?nUU6U>dD1BTtNy,$_Fm?n吽łze]4Jg{=ƮmTmzԯUX: ^n^7˷ -m_WT֞3i`餙NJ 4.elOf_skK?;G2|oeCRJaޔأh7)({{,Gji"hTGHMe.s[ʺ ?mRc5z%f.wWyp?חհdC?+2P% $~Fۚ-ltZp-iZ|\K/=l0Y gVJ=Gv=EU#v@r'scZi*~Z ƇdLJ^9_hD"{^;&T3X "lu0p'h_ɧ|}51rm6#ȨD`ky`tk!7#_f&>Vc cdcÂ0ekݒXTH3$C4ʊ㽬xY5>(9o7GMUEzr7NTRZ<'JIIjc9$eNJnuľU2)ŽjivJcP^i!ԓŢKm,Q ׄOQdO]ܔ42f;xp+]0ά+(IPkCuYv ӌT"yˌwh|z`rJcϨ-OXݫ N6` B`:^vNq2Qh? QLnWˮTb$sјe&6b ]c":i6b+֣(MS<*cq\$vl t[ cf2b@fPh6p.F)m12&d]g M+(ӉPi8%;J(&J +`WԖpYV Z$d BXG H rJc5BʂA M1v#]7cEfJW5gAGE"I ӄ!6K$co0f|) TMk5TX7Pr QtGdU P@S1 Dt0GQF5ͤ3;C%4Vkx5{RPRqe+%cL#r 9 իf!1&伖( BD jJc`V.2@b0۽@VQ=j+`"Кe8M2yoprF뗃ukᗢH+fD' G1!Eg,ipD9!"MȾfΰdY9XѣF/%ܢ]-zDw!t$MY+R/l `96}576O[0V%SdHW=$V*Xȇ TPPj7fRcA39zjN(dUʫ k2aqc56zOy 8eYLZ[]hQx nGep6"YԏD?/A_EYьp6YK1e8Tx*oc]@,M$@S*#h6F2= R?=Ť<$Jxҗ9t\#TϺ^H@'SAʀv1JLUm!I1M(W,pZ{6/@(Т pd..h]qcQm\g$0SZ(ͮJPHƁgUkUkPaRB]$IY6 NX)h^ 4ݮ1V 3li`F+fVo- Rti[ y/GѠMJw a:J؀$`>%U SaCjh\7:hn-atYg_6j\muL몑ԭw@7ۭId3 .-zLa-;-,:6Lk*%'ՓF(=N`ʳㄙ?6"}IaFI*ျ!(A/Q['/sE5* t#"B[uzҠD+Y.T@=ʀ Bjp*1J[$=|"2@zk`}V="Ö+Y|]/w؊PD4cҀ'7z&jsD Eɟt7/`\83$ZT1"&7cbw ;^sͤkP\GJFdUu19Mk.,8R;kҪTg jtϤy_pR 9'k3Dv *UaLMtVk LW Ƭsp$k 7VP8i* HFWP ̀zreRt_H O30%7aFF,Sat9 LPrm+iҭ_O?p4D7gSMhͰ\"/zdYxY 4IzdS((`jI*\ m\!wrt 0@57^/wy<۸[PL&gFz370]1S?}e΀v̸gjf ʙTכd7ٴc=X{1&AU,^Jh@>|%P4J'w@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X JȇF="% <5(Z/@OQ ӎ@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X Jz\J #G s|4J Xk3JgX @/X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@zJ FQ\x@1)*=X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@z:J[7{jJg=׷MCo |^@_\X-`;c3CNAe#'Zg/BA[)9`jAPwĥP]FzX^|5D]_h}:Oo}eaj{j4:/3ԔYG*fyqV oIFOo|_<QQPE Q8QV&_%Ojցy[EkŰ.I4;Y΀je`^K%\EuAo!ZU)'Zh4Φ輘0&3q6տa68uXM ZfŠ(|1@5h\3ZhA$ S$[VO3RwVKD]\r9.kZAr~_1Omv:ݒzT҅Cܡ?i|R뽘V|$uQҽʜtڨEͨs(jAP("ݖ8&Iv~qic᪼qNt?'okg]>gF /'ms&e VL-W t-Q}l-D p$ =G]vqQZ+"ǯ3\C \i?MxD[*$KͷJV[Ъ\Kv .<`lVLtL!9qtȧzZQ/>>2>2|dOߦ'Z7CwEռm 9f&18 w썹C R7/;@O7MJT1cV3ӍEͬM5n@̾߱>S垶<ﺣ-& mCjYJ45P26܁ܭ8-*`)QT܌JϦiU<7G## ;-Eo6 k  1甬-&RRRXHNE:> |5i2uO۸\UI璲4j(Rt:s,|([QZW*M#`Lի[&j lL>R4t^'Ҵt5\K,ѵ\Ę+mE梄H!!s:t媟y&F˼??/]koJZ؂? _̏zAb+yIxzfu mxob;mbd$^r$cdl5>LKam吣04Tɗwڍ9?_srQ~˓u\ ڍj]FZYC_Cy@gyUz{A|e~W/W߼~Я_ϋ/yy+( c&& `GCn?t4S .f0Yo&qjSzB;/:֯m*Kp׃njzFAҢҳo_F櫃9`ZerR ncOٺ-L9Ƶ},q'd~GdΓTuN[.wbAO%ih7E\xگ0$_UV2O.ST\o}G6D#% GZO# NUߪ ,0$uTd>'i}ݝGߦIS[w>O뎋F[w.f{0 o;;=N޵6r2XC`Cb;~J$Ci3߯9CQRC-w=5=]_UcJɴB0 R+у;c,0MɌ%ØəE=#yv[0]RicOK!;,}v=9,D˺< %rK-w4S65Sfh;ҿf$E{ᘳFKnש>zQ#Q1*87nE`BU8$wƦ$,`kA7D  dU= ^B BLdalJP{S i[9uz7>|!B;78?d[R"6Yq}k7("\{6 NR[z39΄FSzV2Y&% KF"v!02Ϣbst&U0c:ʹxqq<$(`M59OKbe2;tf? 3KYE Qtf=ZLƨ"_t$A(iV`ā?h3d:Q=ainfn;m~]Z82RK^R h wÏdfKRmk1ͩZt\v\߼k179O*Ɋ、X%<L? `J襜l7>j(JR_2ZֳT +f/lIlАj{Y،hvQ܌XjaX[9 nvVt0$l7 Oknv8a=NfmJχ%Fy3Ch>CCOAond&ڞȞRB<1pi}n VzP ]4j)_5H!x!ֈ1Zb|ӎv%O+.9Ҳ>YHͶ8CS#Κ䢚d'Vc&k3ynx6EU\t2F!z@8/z88/l5=sR=<'e,$2f2fҝộ֜K$QnGb4 XPo߫5.I8'8wK/7xt~?/4 ߽OCg77#_ySR2UqXTzT; zx6&9qoՎ~럆߷i=V=LGzPog^U\3Դ:<3-tM񏔃y{ їQ2D$e`.bgB.ĝMM <]$F 7骱Fa쓓6y9[&人5W]`|qhL]cVM&ٓ2{5]IA}9@ɧq~cȺ%HМypŸXӥGlyB 4^}>aOAA??)Cbt`*OB *òKj`rTw xxDhs^CE/_XR ([AsS'o=-Su^~םTSS_|WU?x@/V{/`MKV"OSz S3 DDJ~LG"Vt3IMT1A[Rm ò/&,)Wn`v^;s >'GЮ/(KP t<2R:JA{1nmL2e}Q[G~%/ù sw"FB A|crӤjOK~ɞ#cVQ1^ Q" t<2< t{~Մ(öG\x`ߐLޞt g8#bn5; r^lzwߍ3%_N{.%+ü.J)b;bEC6@TDIcsjD18cP@A <",)i$K,OnWGI+7‭iU"bK(F[RvPBRFhNPƞ!nol!;ZHvayVʁ-5o}Γ:m2A4%9f]͠!ɠ9ՏWjowӆ7 嬼EAS$ !YjbMi4FQ"<vڐˮş>g=++>B:G6xegxeuoOZs'CjQ`/\.=oc{L,c0^}SaC'a Q͍QmQ'(:^/I>|+h:S՛2nG#0B(ٿɽ )-BMNVMJe{g2jK}FxJ. M[ 63CtH#}{@>#t=Qw5աd7cAR#43p#| p M: _L&~{Ide']b۲=DLTqY) ">V( NScƌ1F1@K ud^|z[;dƤZ~gD [ZՅJGRtKuO:w)"t֖^`W&?REKqBq8>]ai8dx%(>% St֍kC@!N;7`kkz}`79T+6^WR[ERcibp<=þ!ߌ``<_8 ; sLq4" ʴ9(F8myh6ق^5EQ_n`MB_:MrȺ_uQefդ﯊zRMgś#_?+N_B8}W9U҈z2֫~ūw%^Nk&eM9Fn9k<'C`7UX]L]A^ | 5LXކwM~pqn͓(oߐ,iw͕?\w0m?^hۊm |MG͌}z?G'ET/A= iy&٬j؅?)@q5.ϧȼ_7N5JD`4'^LCe*<0uݠu=WlO5.4J I@Ykc U'S16)dSQ1W@ZYs6CQ ң}نBAO=y!c'&3M˺d;}knu<9x&eD6J֊(D:HQ*[cK^(y}G!cV,h2)Z%-3cCf RZ{rV"1i\y^Q#b&>7~hpyLS4f0Xo4wuC?>.DVi0/1Q۰qUL[( Ԇ乥iteSi^Rx0H2(}@)b9fw(vVsR(WmTeYPHPUÉ*&-A*MAd1rY,"}ۧL.l2+Q"K@+ jlf 6QZC9iU!&k֜DǗǕD5F_1--^~MA"8Ϥ2'z<Ӷ-&i14*)01h%b*ŢH@XƠ sMw^yofJ)6  /d1=]w߳WM93ֱ3·\dCI=y]6n߯~_E޸w+'LӥHkhjt=j[+YI+Fl.!~tkYw+Ǐo2b A=Z_Mޚwρ>]0~T 0pS1d*Ey\.D0=yz}#AP֊VS*`k1ET˖cTئeB3Pcx_|7~g7 C{JM܅/n/>Twa/" lx2Dd 1I68 $'Z;H鶵t|B M@6IƂIa jq:S$BZ̭كYǗeű^^둝}k8Ŕh>+zf2K֧/C>|>Յ<#4aإ|qaM>[T,OqU˅$O "oőJnҞP*gN_WC zc3̎QZS`B+=cN!F/!C3XDQ^i֜- !h± t>+ YN2rLS  }TGV/r\` hV#!ұ_e %NJ[h6GuɲV~<]@վk@͔WN+d=Fӽk8댰M^Uq ݏIurrck?~*LF-㬅9ƚ;?+㫦R lBDʐrʷkV!KR9\de(aBFP]wƚtz ֑Ys6GvXY,lfB8>4.ZTv"5fi'?&i?; GlCtJ,8heL"Ĥ&"*|$BEf_Զb6T:FU Ar&:Aq(8`d+Hj/GybKnfFmh=}m>&LQ$t C%AQ!e! hy2)$ݥ{bJ D1i5Dr&*!Z&=1& 7kx8%7k`<>v8"bCTw2&z,ARl߮H)Ec]խٖ V4" 07&k L`yp92$GNfoK̤@rcj-0ݬ9[| qq͝:YNqѾq\}\cotJ'~Bd1DDq 6G9TP5k%!bYǮP{r<<@āM~&?&i:Gn\ <ড়9Me=nN[f?-íX՗ayg_&a0z ;E~3f7aj-0ѷBތ'_Mk>=6`I:3ulKU}:3jl߯_ 7u4I-$wi<js:|ɿVD)y'ih9ogoM3M1`6鿝0o7G3kGY}_0 fy_ơA- >sZoɡES-`4}oe;n+FGs35^~mp+!w W1tKQs3($yp6 oSb [r: st=p"h!c^g䨧 G 2J {ԂNfi&k癧3|u}[J)ΪO1^֟L}:ׯ녰V#q+|~ң}$E[tB#ĪgD]tD*^Dq(ګ֋[xwyuI9ŝfRmT#(KGUQ9yk6:g7Vwn@[7|?uS y*'#ഽgUûYNYf&Ki_GmW7؁8 ]uGхȐ3\5i|Yf_/oZ>۳m}׳qE|fIe>K|D R`g.3~Ci;Aed[)YmTg[ީ-5lj$nJm@,6"`$9"Z[!H,RDʌ DUbŤ-* TJdmoUhĀܾm2)ypʒȭS5gKE4=UE\I{i-HO.|g7gN?z1RedZ#`*L$Ȫ Q9b[4e1"SPɀ>xL]pIۀb r,ZʰYs *嘜cb|`{@ Y)3c<[!9!`}B+'t 2T~,1cr?aXMJsIC2J2QαuÞ2 %:]6;%^pg3i's$rw$.w8CݙL%-u0;F%&dY{]LVZۛا1D5M{ms|:{3})sVP yB_bx5}t-2{r:։4Ulܾ_*rqqy8آd#BZCZl"umG] u7\mt X+$둒) Ρp 3&I:l2O"):39Y=.Zm}{0f V9Lɜ8>.ҭif=ޭ?x7mnz/}>Ηgqb/{KD}T`p7,0YlgPdߏz'7QZu&,?EI'Ƿ>_d8w(T*❖Wv!l`=Qp¨}> dGG5ctQp>8 pA¡EF )n˖c=1M7e%B1>DmLMJK[@z+&;o6#gǓrRw1~G >},woَvm <.\Nmֶ\?NN HIt̅Z32)UNB:*ƐW3lR5b<:)y DX\PĪo ,T%嵛o9dzg+_,s5S̋ndFr+sj EX֩N0v0dH#k3@ |$-eA/y5yNkj̦Ar,s&_dlnE@d:<T% d9{,PmPVy([^iFΎ^7J ~jMkʕ\o Y(o@ ,PkTm$҂6e (>)Nk_e %BRM¥ƂڌN*M,I'HxYf|+7g^yCk4\g[bzmGUܙ'g}鸶{ŵzt[ܽ<{{wg Fc);fz}ͩ&9-ERRsD,JMFeKRYrQc 2LHقUPXI֌yX3UfX]hBsЅϪ W{]Vx)ޝpB4-vM6OAƿ /\cTbȊ(G%3d DJ*F$ d/`"!ddEM5rh BB(jSYLVG3JV+%Q ]c݌5v݌;ƃ>Xq C bBVN:Ũ *y)n'*b,-C@1)B5DK GPR,{yX~yn*}ш5ֈ> TO:z(")2R*Ǡ"}xtYfS,]SȂ'EؘV0Qz&,Y@ҖIK* 01N0݌?zq-k|͸QzѼ^xЋqV)!j^"O,orݢ̡ʄQ+Q )zq_a3x>pgPa£^Mgtޯz?m9 G=>w~4яFImy#j ڳ}isW5 Bȳ0Jno %%|E͠m4_e_ HΩ=&0b* 5{֑R#9(=:XT%f:tʽ&x}|^Z'TyۣBrg/[Ae-#*BGψˑIK3$%zC1cr uf۱U3þ8 >q)Ӯ(R u蘘|d\Pђ,eY)mS9 RI0#8]߱$t腲)b8"tFR"xXΚr/@zFQT N$UQ`Z" Bo^sr+N+D-)ufx"KAObL$V62W১] c,$CK&o#[Z6]3Š:ZL4;'i>F&Zzp#Ęt2y-3yE9C 5^(.Aq4QS]^˶a3طbO0wBj1[W>z</\Qas| cdbDo-D6(<| ="L儺ZޜQj">mo߿xx18i e@WLֆT2ɩAB69 %3Ĝ5$ ceifX_rZζpo-cr>e]57$j2 lM}eV٢ͷP$[:]Eݕ ߖ({SRO)y@v[j ؐeXfrd]l~)R-w^il[ZؘaRd'NYzm Y J"nFΎ)Vt\=XnE%9(d|U.R pT@X|!tBHŧ"D楘_uKgӚֲL:VPI%hTIĺтlK]&y X"( U5ssFm.=}< ŌF"V}lǂ*fg|3لZe DhQ$>i2/B`Z?OFkii h1:!4) $Ζ5Ae H UN = c (ixJd=OT˿2@9-[Ωe;Js q\?}k<pFA幓›4&ThBKVޟһڨ"$0c]gJu),'wCSBk(lj3}1n!ABEtS:gWG#:򯳂f*ɳx&!(Y; &6덂'dNSr'h47q2AGuW%Yvn_^8_+HQڥ}Ӎ)'f %5ڔF;Z̛WT_tul1,߇r6|{:YgvAŒ6rW|yF=ѷZ-#1f:6G:6>íìkLqx,֕|<:ZNzpvͱmUG]Qܫ@ףNAOZJH0i8ôZY{R߿TSD]-]rf2*VQWl-ם袧OBpz2^8 gl|Uw>$I Ht/FiK|NϛkI2[ >`c1\*|a>e h.`8एpX=fwӹu5`gYu4;B**Zgڞ<>T`߷ +)~oݰ¾C-)Sߏn-VQMJ9xiTkP)U$`Ѻ(ҤR1|&ZO-emzi@Ӧ5R_'=16JZ$H`4d B9 Z8_V U:l2VGf\;gږu ~טm NYzn9`O]@+si25NwRXT`)R_e)U=BZ2}-iB EI&mϭ|%zgD~$+%U fb}]Pbk!0s~ ۣ7/:O_mbȋN Aw:IΟҙt'AItb]hyFa;4iQ? (gx l#$*6 ~ec&f|<\#Xt\O}Ŧ=7ްWyj'Y`2T5X{ 0}j҅a .tTtPVEAܲ^*9bˆsjt46.g7.!-7kE;KYa})}>n,kc\y/i1n] kα?6v_iuǭyMtLkf!"7wznpUKlZW#]x~R7s53zo=!s+\鮡y[\ql1}qx}"CiD6c+OI//NΘQ:M Io~]i]FDžLdǟtCOGbW-^W/[|zʹ%_|Ƣ(q5`Èod&NzdZMǿ\7|ߠG}?-Y} iޡmH4Cl?#߰?K?r&7XA'0cD3\̓BgލMZPq/,|9camԆg-AJ B}0fnJY ? 韓C:e轰VǷ!46 Ynf2y.PY؍Qj5A'ygK]Zy?Ufm5ڜZ&H.KE%$%J<3!╬*R(a,* ԇ\K_{w{_J&vTMgMv,1%g%77n[HڠK{?430~xI'u0d`gB5oG{q8'mڭL]7`kJʐ(v0ŀ-3@:I(*B䞎"wR̕fz.ULyXޮ\_ \Ⱦ,G:ɫQjKn`bbunj5қۋu8ݺ);(SZ=`8bv֖Fikpsc`|zsYo\ކAgY]Fڮ,J͚ NL'ǦfZKˁS:BZ!ۇ^Wb S37&>$ԋ涓EI-G4#-[8Ak7EB6Ưս9AM|i[*0ւ'k[ ߍk\ ԓ&<1/#Q'9 %  q ^Dyh ^p*c}UЈl˻0 x>o`%ay*&lnF#lE=RVooz\{5wg.vZj)3p˳~]~UJQkXEfBiβ'[mtT=kK}L0o>m/}n6 e<[{TK~zMxw;wl4t݌BoG9/ B u!wb:G/X+1˪OX5y|7ܚXuiE赭U2 et<)L00~AjmiijAM"0/XcMis;ʿn[$O&p/3Gi6ѝm^[gGcfe|PIscmЎH-ز:o󹑵ւ[(L+oS.٨vg)0d&S-2Mҍ?m8۠g%._4_8󂑑є*OҦXMr_Eֻ2+3nhȟ T!5(zan8}=ggaXWٍ|L֧e[hČ@VmhNl~16!(Ɂ=C%:{UF)qxu6/A9E6:$u5syU?9#q\{)YBC  { :@=`_y\ZKWs.%BΥțܔvƧS%`v]߬.A.M)_V;o]J/YH+%HEH T!\dpKRbsRW,.hEt*]?AOfGZm39/a%[!,o.j?j+ez*W bO2]4ǖBm"تVfߎ؋޵ uK5[sIeٺzlv19k=} Sά`N2ۥg^Lت5媚ZÄ/ !l+o5]Qb֨R5jFӎ09Unf097)9G55-r\,77ZT 4ZaѵZHv=:{{&\R3U#IյP1z2Giĸ):앩;흔#dviJf\X ;ON@(&Yd酲(<+gQH᥯p,'i3`50BzK0dWu.ėz!Z#\"4 @EEEz2J@b U"V =#} p }ށ[r9^(dPG-CciǦ%7Q2\v|T8 0_펬0~uu՜,H/ $ iFP?ՏFh Sh 0M~/)PL~h@s76 ` p$8\:jF+ay qY2m\`}N/8p(GՑ }3ㅠ(SS.9hWO;>INHՏ@pL~/daa[^ptܬR]שg[Ѕڋe4l:\>%$^#Apgqv)Sir{YBe0w5y=[)dt<v^M險#Bp]JhU'?,.Y>5:sfe΢bXdAbqSsL6)K.:/9~'"s83< 8Cc@ѐBV?P:Q ӯs̉M^asau ;>ZG1zcq]>y@ vj:c{ϸw17{'&(^ b\49ًdGєX&׏ :EY3p,nE %Bs;3/t]%?a f'qlIY~d*hUÔlDnci\ 0B=qfèG=dVZ^8^B?aB30pj$И:;rh_3) +1VF4x NbƺSrraWG8@xl:3MOжltt>ch*k>oI IP,%D4JYVdXid}1\DtlȅI` kQ}\H݋ovY݊me'9ȅdX589*E<.kQRͤLUXj]R0PҲ:}@z^62~|aR]XG͟~<Գo*xzsβy{^=~1Ƞ<6jF L rs-)YU!0A,^2kuThVq ௄p:C䎞[.O\ĶhѶO]4ʑP Qc|s, AL csL&E+'*Hcfo+x{(Q  YG fbyJ]:&6ˆF4Ï:fzBV59DK-)I(*? օߔڍw4r)̰L :yghcrLfH~x/q''1,M;U"ݑDvtc2l6~pJ1.O$ iL>h1-F73lkNb0A/~z~ȴs`fy&cMPΓKAoY0 0Y:֋'^qPL&ǵ }/ dj)Ԧ[uhs"i'`KAB9&:`rc4y1]0evB~Ȋĺ{a^(wKGhL958 !jDufV)11^rIt^7rULEAdqMs$\ $*]jQY˜ [vtYTŴckW_I Р];#aOk~ܔ,SXJD1 0{+W{L> 3]t\oM~jv>}S-v7ގAF(yzm*NM'' WͱiOv޷1/L+\f>1CŠ t5fU|bPI^`5'_ LﳛdV$U1Ov<:V/_gjZ}D%k C0%Kѧ@ \.1GG٧?2`O3WZ{5ef˄ "RDҶ %E__)G+1,yR37^+2{zhΈ>z],6wuG ҿhlXu؞&s7oվ6t݌Ӭ(I//_S`cs?5uŭW}Q:(! #ySisoej'Or23Rҝ.nJɹ3ޕ57r#OE ?8~؇ݘeE5/(J_) ţ@]=-R]Lq> '^3`Gb8g(:# !T<"pkRڵR+WK\LF)*Oa= {o*]TYS{BP~h81ϫq6qѣ2f(W"rY52~P㿂/ߐu%1%KzCJ:KE+d+ fpVZG֤#IZ bQ̳o/X7g]dx$|r%mH2Tݒh1V2_ݒEHZ]^lk i8_zL., q &(˭c."7˿nh'~ܰ]U%zVfam2 .Jl7"RhN JB-˲,3iʂ .?_BZΫ>~*ݢi.7}h~xƂMdz7xOO?T/Pˣw +Ru9ג^ΰ>fgXX`<+ ^H gZ,2\iLhz&S[{ TUke_֯ȭΖp%#cb˽_e١pyrVFnaU,mO#4mcnlOpv)y\fWFp_vF+WUMV+z6 y՘5~PNz3K:jNNo}k~՞;͖<ݑI> ⽤tK m_@zY'qYuJeI7=ktc߶v ;5:vڧPZײ ?_Lozi`~/`O%BVܳl]|t9XNƬ}LUCm'Stg4eSvZ"]ߋw.%ߴ-~۳Y[v^ и;ޣp6=i*lm[6'.n̓8T]ڬ-r 3#_}},Ahr^o_{~q|^ڳh'7~|414+ns7޿EZh45*Ml<cl1"{+"fHZY Ơ/ *2P YR, WOf0$;[zw\5Ce97C)]UU.u4~g˔O*hR-vUQ.u;bP Q߀'[o B2 $htjZE dW(d(MU8/ӴhhUfAZզ@_b= ;JTh_b-.`l`:kFM!@fJr&dm])YfrI0"vkЗHeQMe!8JPeoMiI9RO~ "S1A@Vq6%ҙմ(KʹJGIlp' @-ͱU)PHyRpැ2 mWo#  `Ax4q\\v ru :&tXLť&"0q;?U(hJ)'?ɺ_҄[GkIǏ Pq.s< ]]0p_M/D$Rc# Nqq&X4feM@@. X/Rs!J 4~,'3ŚǶO@" 6|L0>?zׂעVB!:8& f* eԹF&/A(8Z{[ Q1F#40ʕiN R4Y 8@8WJ )0!YʋT ͏ B:G[ _3$dRCx qx[ժ<6Bƾ~< q8$p.Άod^7N|ZH:Hrd'yr;*dݲk:,?@"W\ҔMg^]}O\r3/[Űvo(XltBHG: Õq"F'j/&NE鲨br 0ӈA`K{ A BxcIP==J7NqZE~[6!o,\ O N}9dTK/!y&ARXGev}O,(g}┲nTF^jCT!ӛ19ҹqK3 e$ܯm;$Yk% UY,L0ۖ`pdw3 %rxMbCkȥB"J.|NsDpO?-F&>xxHl/CJb--bJn= 2Ot'TC HjJ -&B!D tj2A*ٯP"eos! DPBHƀ:"PXr#s^z (Ye%PUz?/A R\K.(<([(HQ.,Ѧ Ve6$ƆHa*陶'xl!(ЃVV:b#UhQ%=e=.3$uEPdHKء>#{>YQRp+qJ_դWm%NyQaP-`kΘU3>9?T _~mrSotc RUyS4#92AHb ByOҿ>iS8'?;*ڼ=b-6ݷܛz9Ƹd1]ub]gq\!U¥B3^?~ɳuڼk';)AJtnq6{EJOl?: {c lv;}ՕS˰XYH,2zwzcOpZ ^D*aj, kCKr6d:DQm޲Q)+hwlb9RRvIcOHJb{vW(L.{=K7n 7>5z;w{Zl׫)$D^w,! zm>يw{q|ԗ 8pW]bտ/Fz}rbMq{z_bJ)C0'o37{# b҈v?1dI1bT2`({ 1y+4aїW] 5I+EI  %.MڵEݧuU$CsY,)]Q;O眗j@+L L4FmGyv;\G]E nhŻCn1[g8ĭ+ k0# 5׳}hAun4u,0rBs8U{iuW|$EmI/U<^u s9ѷjXn?fè'$5Z)(PIFdkS +f_ɱrYZROf/ۼyTxv'i驶l\r~>[&[ww y+$) ,:A2ڈ}ohŭk 㚔 òw4vL8ޞR8uD6Sdp*ɥ ˜- GNh5Do& `S覉ڥvN .ҩrraLYoT(Ƹ&EX4cB)|_<3ENQ EYP:{sҹT9A&RZ`TyK>_q΢puDtj,Dmdu?~L ySQV6HkQ+i k'25༕WuY\(Q~{o D$5aBcRr7+kF M#O h`P) >ŕy^5Tdnf@ ijjk f20K^5',RdoԚKEusc0 C^v_mt2!t06,n^ڕ).R%z5b^] W1lGK%|0po- ;8@@ӌRd]^!+jMSM}sC=6fii헏'Qgy)TJqrPBM*YY/ҷURhS dm;(H)@x>@iL Hڨ?4w9/͟۾yɗ!+HΠø(]/::M-VA`4wo`7!YM `$ Ly"˾ VHĸn5DVntoeU+.}n'އ) _VW3#ij[ZZEo˫5RToL(6Tm5 _b՞|hͻ)l" K=ډi, /¶i 61i8hop)G-kGo26&-FtoLP(nۅܾcKT!XmО~*ISR5.arHLG "@wo)4|GבPjmPRQG<;=me24&lWw5%mɠOy?xS N1iS_Pa5ZlfgQaʗ5QjOqfhGyBmz';,X+S"uW)tzkt(dvrZ)匚ҿ ;fh|S0xDlb|6.Fx]e0tD\ N'l7Yv?zu?N2<KE*]Mq2/0*Nm9&#[g91}ɗ˶h8f?KJ4M,1?l43u{|_x@$aPWyYK]#5 {-Ma\~X{/ ܱjg)I\^P8)z{d ]Gn8{ R;mʸTB;tW?bbBϲ5u>պ+ fBcF]޹VF #:#vR/] NiL ~]ꤚ/TYyQ7i^IVy>@|fco?Rzm]1E~i\Jyok/Y^BMLrS0:)!& XBQDLuⅬS5c_C1|U*f)D3`ZPjgD  =: +#D!ݳۧJ[ޥL9 xB3w\~ٔnj02,nI gFFtw' Y:@ycv cmH ")S@KGH/ZW1PaK4Oy3+B;0-]vp ގc|Hap$iHqG5S+0QTjD!$cfb7QU G>ZA$3ZfS Jkw?X@'@:vVssK-Gl`C-g?7^qR"mPi 8G8c|RF``jou>)H=ht%ML 1b#Ig*s`zGW^@j?RZ>o6Y-%xp4Ѭ}HQ# !T=uBb[;]ZnmRr]j: bLKq^ˉ6!ʩ8 Ք*U-fL U_>B%>~pr iv l9emWd{4EuQqߨCR'-XLHÙ=KҘ$Ҫ\y-0RzGJRTޥ cx.Jڼ6 ;i (/, h^SYy؇b\UwTzj K@bEcnXI4j2REW7`RWa)ۢ aFxH,egG߇diTи4CB>\쨵{µ$؛8ߨJI5sʂ3J0 ![ aUFH`*% ZOR%kiqGrOzYDȱΧo=jPke%vVPRHM vx;%mClUUPUrv`~բnF*#O,5( VVzY/XAկ/>ve8h m܆Buwjn4t:[Ax'05qRP`IEێ)1Zk gJAqITn2Õt[N pk.~IY ҷE%L%3Hgom2}6X99s(+{ֆRi~ ?؉4.FԔQv `oME2j*h|ﶀg{0S |a@9O.8P x(qd8Ļ&:2{lxp_&-O8T Xɽ+–k^YGv&muteV%ӅtHц|!jkʂҔ:4ZHֿ{֊z7?o_h2eÚӜm]?Jtz>TevV,;FBeɃ3UsKCLy& B+R,-BKJ ~ ڷI_H٣{g6oXwwX4HJ˦|>|2|Sv{ѷ?i01':n :d$4K )ۿ-d~f$ kM(6$6WcD T$ sz lo|Ї 7ӯyD7s-n_.M$JdBbGް&K+t#wե*ƦJ%QN|/B=^'zq]JZzbOhU6R*lֲ!Y݉? 7f|[C,^솽" ~|-D“PP$h=""NiA@%MNp<1)Y=-R(v !UPImGgGޞ8o ȓjc4)8r*+Bo)N$'FFĥ²ٮmQFfq*b)`ժjW; 7{z*>%AV 瀱 u,e?Q.ujl 5x~LD=/r5"o2[4B( 6եv9zF-X#LL1-v>E))O긅 (C$pG0 Z56IEBt_4XH_GlBHo.7LdrTЈJ oI`"EG,YgKa,1cb5:hTyiwD6H+m$9Гl} 0^`a T%K,vS݆FJTwޗS3sV/䲞1Q,G|wzA0WRm/m`H'`:Df1+R+U2q,X7W6^@8}1.˧ڊՆezߋ'w`)169<OIs!'Ή9u[؏ Թ? W/c>eS2oʱ`$K$}?Vx6. RqI{w|B`E-naN vaGH6Rd+l|l]wDZ%:T85nHUeAQG.uT톈`BOHq!)W_Bh:޹:Z ]nr߆M+.=b:ANnJJv'Lұtlʵ=ۋK\HB@kmNk߄+wcpb(Q(!xa b= 6U (8YpI6Y-mlWxXP2I,_AHL >ˬk{`VK6ƸǦ%oCLjd9|٢'wΟC|ipLIWM߅lݏj$3J)_^iRQF9&GI;Uf=MAJ/֢YaF+w~ʼnohf9:isT;I:D蹻a bVYTP &11+bQiЊ~%03.xSM%V~C34nu`gb9f?|Q%}+)GXa,Ӆ K|٤CjUa,b̉&8;tnkTRnH[\zIQ_9wV3?6)VQ.OVƠjFך,vȶt_01bc~ MG{(F6/PT.þ N &1_]>S]`]ξ MuJ ps/)\.GFA|/ G%"{snf\ߗ0|Snf@*X\ré,I# Dl%nW|# ū%VH [VMWӭ$ӭ7K$۴5D6 r$۷S_J# ,?Ѣc{y>k+:,DO:Ȣ}pz+k<*%%+l9>QZUnB#>+}D59Q5zN6Hiڮ&H ռV~bTC:ЫяOcz A {Húi@»5ELI{KL{B'/ZLm('P~J TNhڐAށD"PϺ}h`4!( %q47(dF23\xO$RF\J(D|+ 턥J#d'&lݛӘƚP_$!O@\`spKqXziRN3X2 ;$(gw]ΈE6SMbPn9R2쫟0. 9whr [)i œJǼuyHjNj/gQEvVJ.̘zYfCڅ2չzs Q1 +2f (k=.gw;\ٗ6r0n}~1p#8!']TqЧ*w2Cf g V!d_TS3 Ƚ76!{`+Y56m]fkfGΦY5ˮQcgա?v #~1Տ5DA4 YL (Fx^MFrȖϿKMfgDo9 .5|1#HPSygeBjg:TeuNJ%hR;7Jw}X0`]^x<}/&YaLꄞRH=Thѕ/TlIuX~:d2ӴƗ/m>~?FFpE}"ܳ:nz0,kzO?H$cUrJBao "'IEMpW` >P[3MroIO?As O* f|O)VwkY~6BYය+??r5 |l13∭zGTyzf,wKQn`30α`$DV6J0Fa6YFBZich(xQC{e M C)CsBLG|34F$^8m*PS,6mZ|`3DJRs W)يL-yᩎDn;=VӜ;%ԪOjX" JDgL`EQʉ|k V)ZơQn+21muW4jfi΍(<$}<>i5?^p2%+l7('M}OjiG76pIr85ҽ)e+lYYVդ4ǫL- amuȭKѶυʹ cR\J| X+l8iK*~0hNHKڈCpym͏gΕRi0;uN>)F/H6(4TIr!. ' |dίtA _ p)vI{]ܘFA{?^GW+myhw(#c( MWu۷i9 S9g/&spl .F0d0dBgm=%2K$E aWi4 sc3M 3czA˔HMB % ss(V45[plx?5)A^Ij8C.~jPL\)*OY[yPbVn+g_n_ ed kJo+lO ʝ:o ̡ gG! 0"0ɠN_ǁ_o`>y~TdqX &pj ^k$]Vaa~Qۦ_]\ap%tp`R;pتLwOw96˿ҍs)a@&v0Ŕ+8/ph\:!E!{x:78WO~y/NޓD `il+m> ~<.XJwT"L۬Of"O<G8?I%#JGfُ鷐~,*䅻n\U _l0q᯳f3\]t?sǁAt6D Ұ*wn<߆N$,kOVz}ϊ/n7TW'yqmu雏ޓ0oQ;Mpn9f"$Yաn㤻j='Wb^ԈZ%A֣4 d1`d{Y~赝!IzV G%ÿvPDsҾTԭ[sKig>Y#;?)_ q⽭%E5*d|D}!I+*5~& HX!㍁ -e;oLumB-+rK,3@e)q5e +n~AIcsIW.EKhXkxS/x93꒝gT`0U #JxeyLw^jiWCi$ϗ[irjǜ`AAW:5,="Y%Tx7Z7MRoVbm"m6Ϝ)3BEEJ$6R앫pY%Iޞ,Įbg!NH_W(\aK(P8BYԓ2;:ޝ3ű\rQ;.=7}Ѻm/OőL黓m A)MF@~vr͸m= j0vq4n.<" hYy}7y5onCۇ?2tT P$H:zO3A%0ADZjo ]৫mLnt1!I\{f-F87ὒLsɿ>|$_>^~?ϫcs+; |xI`mtݴL+ܫE6-p8x{ȏËb<0gq9/}:n):Z`z^ M0x?NiqPSʔ޳_SƭD}뤯Ll#)h;o ՞Oir=ko`^K\B[+mk=줪u%ݳq֣205VҧkYmp9m=/V:/j4Rd;ծٶ_ǹ{.pmvZ81Q)ܰ?}z:ӌ3Yhж5R>O6p1Ì#E_BzsNX2bRTF+Iѥ,0EF. 8yV\`ZzIauI2JMA m[}VԌjJ&AM^/辶F$loj[O͹?ꐦ&"V>c Z7t2+G TBEPպ;1D |:"VY繃4ue -xT-{jLiiN/zj0ƹ-W<~*p3:AqM$X\ #sG=iBb 톎i!RD{wY_};h9$Q) :N#~_*z$ %.be^dRBiPQ2ν)SdDILW˰8Gҫ^]]~ǿ`r?8loXqq"( >7PZ1F<nדYYӏ46mȲ/Jsć;RǧdZ9^)3 +/Q_'$G\FH*V> "7:b %J㍷:~dha<:2Fui;oZVQNx mEB0tꎋΘG!l7||y~2}L37,*@ܺ̅8ENpsjW> &8@;>^\(u'UT1˹/ fϙu\et[g7AQb ⍔]D> ~fxMFJW܉I?33'wEXS:IZK/L7\ 4hx&r VG-I͆ƣs'xٮ)6h4{&DqN8W:ICW:݄vqS3^\g)~o&$RTX? gʩW^O/Ob|_ݬh2A;00Sjxr* xsb2Bƺv79r.l%|լIk͡ u!8=kyl&F_ǣWf wox0f -W(izkb3TGwf)U 341bϖv_q-+櫅nJ}Z"@;+3b2ȬgroJ݋\{KY-7]yRZS{o>ChvVto$E=B3Ll[E.-2 5u:1*-=u.+2pX28T#uKM{*sj .#X "r„ 2AE@@;T^x.A\,WqR-f0lb4CRۻQp,^D0RL{ rBƆNk\k7-=:@'UyNVTv BX5є&zْI3R!cEzx 1LC *N>C%WKiÌnj4+;At%3:XOF5ń$5V Nrp@%Hzt+( ͎sy_#I;)20^c2'[TW.H^]>2J(r's֘Zkƿ {EnƜXixIA>Z^ 4cLÙ4aG"z|.\TX2.-eSdU<=YiMYtu5ۛnV*FHR]s/6aZ)NHO ՜AA_'siΝLxDh8XBѡVڒ %[R>Piݍaeq6;yIRk9lEgS9~:Dg [Jx2k _iHZ>2=+UXCJR#P=>nFɶ$ LWC)jѵya]iT?Fͮl5>OPct}<"(w{<ԥ$4iXX5"Ndu^jz3A/.NAeN3ʍȵ iA>w@ᝒyB`@iyM p HZTPͦ·J:oe ڑ$ Aʴasݵ񠤔Wx=%5td<[f+|е5չLhqc0fzM3C%i"YxYNK4Oӊ3#nJk}Xuݐ7չ-R Qnt}O1>s 0,A<_bpV4'sWcUCBH]X6f ٙiӤ>42^e?uU,bҎadMҧI6oٓG\i 6ςg"atrAI97U17Rara6.̤(#S\de%G,e/b )Π& qn41d19p}@4O#7i?en2S+kTX3Zr,+2tyA 2^Pɝ3d4fSAИMog˻dAz%Fu;g5I*dNd5&]t5̤tZD}Ƚ'2" 3r]is Ǽ1 ,ϯ['4Ptmt{tC!]%Ts2|c PZЂ@V&$!J01sJ+_͵CK(E EOj& 8nW%(b7ѰFLjX!g]5݂7sOU'}mpOje /g:iD dcI+{E_ sN @:2n9棎8Ol(aYe/2#/e0Epuf9bI5QҝvZ*|y3gg.!Pg +g Y$2\Qs-ޕ'Sdյg@UR5{h2O XxD`FTGw"g.dh =c'AmΚX UA(&%b'Q&m%tT{ I9"<6?kʭ1Ǯ>h.UN8 g9M:0w`-h7=*0S=c9yAT}ˏJxPewtUL2*e}6~n2?kvm(! x-p{@S|)Z gJeܷ=L2g=s1 ڈtc ZO}7T{@\}q!G%+"!G\vP9!M -BS}n2]c`RX"g((`qݟM Xy@?{WWm ^$wH{g@h \@H N%pBBթZڅm3MjM:1A|#!kOT,N]1flo`}t"4"%c2Os/G ^F2Ч1qx¼:Y7|*ljO>f89ԨL:XGBA<ܽQoCyKR˄\̑} "K'> >{\ME*dƕVej1t-4nnI W4m5]2>2c]51;ìȕ -5m.:W}֜,Xq<Ư]IըecT;FjϬ"l|˥>AzwV/z5OsU8S~Wz#p'3^.hKߏ`rpKpU%MzsoQEUᕄ>wgoŒBN*OǾ 'o&WABiR"1uPmH"V!Ɵ w]a*!:|~C8G&5%L%n)[f[IǨm֠>iX4xP/杏9BM31U1 Uc$$ [BXC$u_v4UFԛ)İb`hA&W[-$T*j0p)92u|oɇx V#=R<7r4MUǯͣgf}\ƶ%cq(ELjR*:V4aR'h,j)E2)r*^g S%8V΃=x@Q n8DrBUrm;䥶ߪ/kUk$*gYWD~)CDO_~-,%L1Jd$L%hI<vQs 1 iRna2Z#j$J0SPGD59a`J`G86^[fG1rU{2Cw"CGUT2?|y~#ǘy1ڨNULsExnaK)# fwR**uM>R \MSԣ]*Uۄ> c5 h;)h6f9vn<'/Cl['ONy͈ >m' qZTHe51\c*lkB-ˬ/T51vF%yV*|Y_uTd\MRE7]ǭغ|\CΪeF=ćŎ{6SЉzN'&uMfA9՘yh]fjјj cM mz564+h8n=K&{G}ާwdwdl-xrvrZNW㖒yՐɴZ&Gyk74YJ^,$K%xPYdzr>|j>ȵYFi]el()iXJ, S'v+Au" ,_jغW̱cȪ)kF+ !P_]ͬJ|6TWg{Ѳ$ 004W"\Ñ-9̒C82 4L# 4L׳lܼ%0Wn`Cuk̗lXQai{^l_+[V棳sZ5oQbkbpbgHVu˪egB(9N-iw{,\τ1ܧ?>tFʨ"~IK)Mj@^ttf4rxՁ"HAl uʆ,j%XKEnolo=?~i9}>8}qիFyXݙ~ň(jE=1 Cc|||1¼GnK~XUGYtM@%z򢛟n>)٫q*P?-BQϯg'i90"Oyྸzubj-?=_S0IZ銿ڿxyž=kלa`rσCtxnv~Ef?NN.[xמ1V_?6?_񀷞~X"l8/DV?OD=#Em3&0Y&sX~q{}@Jn/h /Ys>k)lY V$.ec -kBs@ y};6@%yŕmoΊw2?(5F]X4w%礡%Ft;Rs'EE؊qw /'_>;e`?Z~뉤3^+>MW%~\}wONYw&TO~}9Pʵ#&v5_ynZ}+nC;ZMMd&q6q,b!yjcߝqU!X~̹]䆹k~O(g7?]d7=GCT6-]0r}7=)b} G݆L5 Vnkk!2?$I]x<35O&KDÛSG}#-7pp0-FliiӴF|r 'Ŗ.tS| y wҥ8ϱ%O,79jM(n4g8=W]3:3†*I ڊ Rԙyk8(qȴh{ P[hZ:֚e;_G"W%-J7 JF֯5-ˠgB_tr3|fMP!C]EA'֓[gl|-l;IAKM1\b ڦaxdN!$&$eN.&X(e6 28SOw 3|;ylbb&̃= ؜3 ,-Aеg irj3îM&Kgw%O;q ÍM0mI|ʕ[NMJVB wU!g}O3kK2~d1ed3NzJ,1>)a^_Y!72~mj$a4RF(ᚲ(:mIj}&#Nlⵙ|ǹ~m%z>8Z J&xM6)_]ҚDLcjUOapTk*-Hx ^jt2D5|3 )wՆ۳/7L'Pʇ@TMx` #k278\{ = tbI5",f9tm Gai n\.k`77A_o~k!cUѭ_7ɠĮe磇}G!UKxhtٻ޶r$Wy +@28E? ^cw9t߷(˲,ˎ#b4ac vT BezLqO =\ŗwtlNh\Y0kKn o~~IrK:h/ b:[@Gpxn`h] H~zXotYk<^" @?ٶKcɚP?iS~On`O7OnO-~RVC?De+bMpޣ*r{<=5VyVWӹ}~. '(oc pl(+yc\2Jpv_5E/t|t׳m 4"cC!/p-}Bp.(9x:}oݪY--Y>ϘGlǪͣ5DH!4x;f(3%nVzˆBbE?ۀP&rf"g7wDpXޝnP q)Y^݂vϚ%xnb;sC1"}OpqxO1>ѽpIJ(z-sU}&uiX~Y0p)}akqc'+p@Gc{VmLdgöO2`:/Sp-}r& TY[Mu:O[@8koNno8wg(D A/l8@<;wA0{͓hZC(21גg; :sW:C[!']l o޺cJ qTBkv %]f8<#~H7ƍ:9 wDL}T7JFGa Tةtc"S}lϱ{Ώ%U3Y]f^lGTd(8<2>Z)ٻ/o(j_N{q<݃Wm V!g__Kv!G q}YbV$d C:g#Zrr=^z"{\y2!kDL} l-V39{$"%CFhE}*^ Ί7MDDD Ɔ@tmXv,Zpr/2G'1{i-fx1&~GE۝GPS^ Rymc^ZelUdU TLAL| J[-g4/=zY\[X~Yb=ӭG]Key)0dzXu6~ݿc9 .ȳowKKnv6A>^ְ.࿮~>k/Wdx^eib#kW/{_ux|xzȎ̃T v[mأYd]kc@/Am@մc[9@Z|_`%-}w1`Xfifxӳ${{>ƔP݆ӓvXož颀JŲ٦vb;L|;-ӷ5?-D6hHGM|-sCo[3vpLm5kHflX88oٷHd0b5bRuY:u͑CDD2AJÇs@A1Ӭ-SK5 2wkWҬp2O]p<~lFЏ߇~>a}#a" >sΔBX `k,&UZEY0lSXT۪wPьOxS=KwEԧ 0!VTAgcREUѢANVuń@PՒzPb+>E/yJɀ!"t ;(>jY5I#q;UxX wth_SŁ U*cKa3kNL׏tE9]Twmkә-⇞^D6N#yH6uVK fCb+5!hD'!桗9rz"`^a(1*RP[BeKwn{FӜ^ Pq(>sTڬKkc{k{.A-Aֈx竕tL_T't񧟬yFGߟ:go_W;bMj8 7HF sRN:@#]S줌6)6k6#dQ]]T-hӐ"Ue` -u}ji"(.0"[*ZVٱ9HZIj Uh-==Joi^RCXrJm±g_@ɠ& `[jw*b\_[:=d\1'BSU%E/&WK6ʑj:O/M+3Fy=)nE^gT6Nt}m%? 8ṬFtmMTszE"s-U;˅R/ U+4#sh)@5 &,Ee(թ!mwB>!csݑͲEH%~Ӣ3G4cbmAhQ{3Ďv!vp }3XՋ*jZ%u+6%ؚňMT/`>uCd׫7 wSmC{ ,^*3L doF KhAAs6{gq٫% M?o'o7ut|pF4÷_ik\ ?l츒Gy0ƔMuL!YPu'>fj.:".W L MDN]Ug8D).QY: s#G"΄ >'Ϊ+JlD2z9m|7jNg7 ^qK@b<1KI؟.5h},Ъt^ӢǗ;(6 "`W,N|%wac6&*nDrG7+EѴ10暀S$Hxw_Gq{HU'C"8D@lW{גݧl{ {oԷOedgpHƖALTU;ErC8Sm3ĚTY !!q[c (U6|)2 *w?[UVõư[osI9Z.F@իDUQ F -}ƏgB:PV~O|r$PSY;qdug؉vƩE!:@*Ohc Ԑ@w1ٔtR8ß`_Ơ+18f>E^2MYQd@"LTUS$(ae9fZ||8`)tQR?3bU*h+;=B$kmv{qTfb;rݣNUi1Uк_e}9K2دKN f$nTVLxbwT(V؆JUE^Q%[}Δ[p3zMı/T*Bp8~ۉc8 cȭLMEi l RڸkX@jo#: 1Lj`#)E T5׭eIs#WE,^kƜC]0;kwoG sL8`E{wܣc $6O-1AXZgt-m}J3 Y71EWa;eb&Jjn$+%"}YR#.%lx_=c==G:~wem$I~[ #2#B@_<4fc,lyeFRT"ɢdY*KJ%"##2K,*De` Ma ңέlEړC!uDpyDznxzWrG=fǤVvn'3qq߻-PW)V~'>o0? ؾz<; bTݲo߿9]|L7Mqa7TU.#ں7~~+pF?u0$@8x<8iUeSUW6G Bͤ Nk>!:`' -{ΑCΜbrZk0RA6ɨZlhCAL;%7&2˪Q~)ɧ$[U q5 N*U(v%V!kLP\'Yo,f PcwrDY6^):tAеF_qe cA6@mbE- h,0FZ_eҋl͹ T>9: 2 ٠hh9o#g )a]bZRցGy!-j&~z# 2)ys8:>?`#|>S/DWpЩ-_.ɱ :v18AaF' FKUNuNRApVRBVc_GrmS寺$rz+,Z9G͵_#s;wzkEԿNE^nRZ9QtuNںw7`j+Ѷ\UڪG?C..)eꮖGDTcY1bq3T-Ofa&PAД)ʔ^OΜeE t#-5d?,jN!3d" B&mZu_?DHSd l}L>6u_>rY!hQo32PESNNwXՅ̶;J}\,Y)1T-)E,lXFԒmXyDZoja]uO418nYus8ābwQ -)#V T+6gEjA_#k4XcR6;ǐ($iEJDя"?y>\ѻ)<|j2MdvQ ܱ%&[!$EeA Qcj 5a]f#gDY՚xAzN#Xu` &z*—,tN”1 v`[;`n /N cCCs͗ Zn$9 RoF=#NM? ox#-ּ$@`VL,vCvyW4;sP#6FpOQ,lQ[t$ ҭlź p?k^d-S@֖B$t,PM >q}/z9D!K1Aٱ2?w'tUub܁ž$P^. 6[IW[WUųUF M-ֆX iD 9A"fCD9&՗ٻxq6|sVPZVC'ŮFT:a1.褷yl sUΜp]RNA@ )aF))1ut@ F PMw7ΉzʼC |5݌nl)e޵Y^~*WO[̹aug9=(TM v仒tk4F+:A,M Ea$/ojRbPIʒn1GQاLA/T<R/F DH)r2HeGQl83}r}ZW ~yfbnܬy9&tBWe?{wRruOy+= jmX4|0p{)+RV=MBQ#sv7D[ID>qgih >$7*F`_|&0H<[TI[ PN)m5qy$/{zsU{j|wWt".Y9jGrrl©dGk#)e%FnYm@}( 0'kNjOjIO*vOM[]eݴ;E$SSSOd>W|r'>.;ք 1KRb!l+HD6+2ekf$ Pk2՘rQcBA|nKo75CF lMނm?܏>COxKу{jU#2#xWR&9yrHvWJaCL~Zr]T:wb%BPm|@>UۖkLՕCb"aY녞 g.M &LdՀU5c;&GS _c6$(‚FW4jycdAGFQZST(A18矼C?qwQwzhU.`t'vs ܸU !P!z y\Xd+KvK9Dڕ1ꅽ21A#WȴӮX]b"2^JOMm0zsD排{Re?"G%7BeO>.b"Z~Ga8cBRۮ=ED1C +`lUE&À%UO[c|KfegnGPߏs+]O GrDa| ~d Dfj +Ur9*̪kzKڮcwqgⲈb\1Dlb1NGC >Ǫ1'&:$҇l e_N̻$ttuRm:%w+%VZjY6TB1kBPVYgXňgO'/fP Q֪ZCQrJsѷdsUeoHc 9ZTΖlp-FU' HA;EWŷӋgyӝ9 ,Z4*zvq:6ux)ӄN1-2t i & XC"PD7LJt^亩6tNpchn*(ШZlWHb(xNu~|t8<ZgYd! FJ.[Kj]t sX]KRDb+"jC~vi&)shlVX`Ӕ .zb]|bmju1Zjg*h_L qCɁ=2ꖬ|MdF}P:4o$ CYNiNB[( $#Yv֪yiJ i玒{ۗ 40o_{\u/rnۿk1F|KԏzǛo5Og* R*oZ^$Y5M\CI[?ѫQ*NѶ*¼\&n#?m k-Ha '(BBs;![}F/6$[}HRI\1 [.(`OԺ}nE#?|e?+ec?,Ǽr!UE!VݴH*Qr/#p azD d-r,ںǼb<SCŋИxTE+jy{\-3Ql/k͇"ϐ&J(ArӜtbfZ8rԔz2*lUCz69B4qCIQ-~3ВUTjZp5l nXJ6[ mmb6Kn Z&f(W;YXh[QmfE iɅ2EeC!5 {ݐBoE`W)4MjkY)ek Z堫eZ傚& u͌2X|(C ϤƵp }>x?[G~yp̕'=K*wu*Yg>ﭝu$igt13f3{ U;1hFs5̅S0ι.U>:*" D"M}z2yi.m|¯LlˇvH^ mщ×}#/oG9W:ښ={ \~,zpvG-"S 2\1p*.}JT)Cҥ'8*3-XlSCZ,nwz/7W £*ZZsZݜU3uk6V^;9"|aY&R[b ?)Xq.u:,2DcW簜uגYKK-=dC"Hfn죾tV :x>4gE:] .Y| ,Kg8y/!WYq;kjA R2b-YQ묨ZuVlY EEuw>ԠjM0C,Q0Ig֊`e{B=?ϔ^:)]qt _0IP.p>]r Dۘ \"E0ޅB.3N2xƧk(u5hs9Vv 4|}הVQƀ` s{-;$*:sL-.yNkiaWHSlF2W GqbmUa 9;k g‚7LA֖L֬p '<:v)@8wb@}c^d q̀3d3#VG){?΄I6b|[h IHxJSJiJUEQ;,ʕBS.^ %29U }&-9"9z-j&*HaFQ"w`b-HE%^ %$b ++ևB)`:p\yu+*2.Lv)ߍ %mV2]6<+聯^~x쿡_kxl1Pchr=ʎGn9h ~E<&xFw6d5*y[yEZ|{4+XL82-̂C;08غ]. & Eݮ`7n8:,+ʟkaҢr!ȳcom%D%ٴ6R(^2jZ_|>X j#Z.f JXP6gDk(,q(3Q ;ƼsD%)HT&%+I5Q\A >mw5\\Su} H  zF pAR­ĘtuoS(\1 ն̣Iݢ D*2͎ &$MݺHz[1DE.)vInBdR^bRzJ&-Pbb=Or{е[ }m7Q9)Tyd&1hܚ(itwoŤd{wo!6OOޭ"viB {TMXDvizD5 |;^5d~̷dR9bGʅ?n+7< j%H35oFCLϾ>dVcYo%('虛ܷ?PyͮQ|jd)ww86,YZͼ55~O7ՏIm) ǤD) i$8D"W|Ȯ jUG50f":Pj|Hv#kZ+ f!>os;n Pus{7s~RSC3AFz8W g;9_WGmV%Ż,^VlzVyB'Wǻ+~6޼(*Ҥ\ȡDAHZZ"Hsj!ێ Xy|BjT\@|9L/.7oV|"t\#IuZԭDB 3H@XI7 %EJP2e j'iSruj%bS7aZsz1zhm3C|c>:"fg汫ƍbůWvzg|0ihˍq+i 7~닆#1Oٴ yK|yVs|bh)9TAeZ2 FYu-'Gi]+j)qD5MRO)v4Vf_q50MPU[tW P__zfn9r~ i%԰3*e ˘9ZQWŕi3Z6(bh!R ȭ0N)7JRŷ𔛜rw>\%YGQSGv![v2ܵkSKHv-:2fizk_U,)h2 DjfﭣAdTPc 5V7DZW;X45jUYk]r"21 ԯrx#4bC(y◹x~Nꌜqv& nwxtFfo@\~]!油SxIFDVa_Xa\2e1ie EphzS9 @m]JqT6Na w_259/hXYV@&duRέw(YK]vɈZQٕAm#X6xU3~9Dxv F%?NZm;^䮻܂KsVlN`g$(RtN6qJC)C$Jȇ+TV{INm0C(c:'w.bE(Jwk+0nŨ#^43Cܫ ޓƟ pwCpvVjb+kNu{JFh@"Z'j"(ut(: y8$PS&q:a%3l #03k(-nN\xf"0YiX2J=g.Giy{{B$ygS|२/j.CP *meR K _nh:EUc %$){D&Tc]?nDTVg8x yG'nIsNѻD8(y j-xnPBnPkF3s-MD{}K|-w;(}k/G<F-"G&ESl9EgEĀAj -X$O5~7AQ"?{OƑ_!e>fȮo}|$Fxz)ͰI,{>8((F=>)9%ܗ4#Vˆ>Qz[NS8p(ppwٿ}Iu pIwF)?v wCoʺq 4s5L؇Q?T:.d+o'}}8)giYGv5 @M^(w5lAy[sn5W8R40NhvI8$<·bC0VlQȍE+FQ- 'ٝ).j2>4[-2(ZZ;,^+`=[ YhĉT~oX˙u`v#4;Z#ve8]}n.'h;]oZ$3Pr< r1ł4q=8UݠVʻe;F.1rY]N)f0aweVTw'Y~/GAu/6 M4VsS`t/o 3EƥIɤ`ZY WX5 ~Z3t>AߗCKAg|?iV ~xOvI3i|PZLBaz˓W:}J>vR Qd]!1ԢY^Pɥc`1X}wÇw'&zDIċ-Ρw5}ƫMzRČn*R9`fTKf/?ߛ7 U nj,:O8l:zo Ol6w,$Ǹ jw6qj#M=u~};;TO@~өewL)ҵ)y{JIg 8a `wTLs,CzM+ɾ ) tѠ{wp+ NGR^jK{ѓm,ZC AȢ'HKH1 .(}t% ./*V'KLvr_Tv `; %8P x;K쥥`엯OZasSǴ^u{^Ы֩Eδe.Z M h} ~ALS!\)K)( `lc}Qu|}^'=hN337g4/+ԯ_{zx<469b^O0]ڷβ!Tx:KwUb(W(Ӫ+x#cR0b8 +07; rLiv~9fP&M5!*37=v> $uMݟ.&Z\sRFȫ߻{hNJ O7o' |%%"xs`\3qZfGE .{]Tr$3cMJ0N=՜b w'N(*HV!X\zF+yJߕ/UL/U?*&gWt e$*Ga ЛDŨy4՜+V LZ*2Fʱ N 0+Py P4f#],WFŌiHo4QAO1.W#>$ou إ?T⣺4\ DM< h 9"Prdg\RikQ VFv8bˢBL *c`/sQ!pCT p LRF+$dS)$ ޒ^非c.`i瀑 S]:.0M2BY7#ȒQ^aCxSSu2䀑kI]falN`S84*c.r {0N 1<w=.ٷOeh.^7'5Qn)s 4A\:pa,CVRZxSz`/\Z@,T-`_#Ԭ,?hR;ZB}6Bӊw+pCd3tGHNity);oAJl0V3cAH=Fax`[qm0"{Y48̐Qy7z=vH ԖV6p%ZmH OR|g8G/"I(FLj1T OH`V )^4D)\/ˡrŰ"o&9E4)8p`$юIZS-2RW;DĜ NPD|$b:0 20lff| e^;Vj5-d2FiE<YQL&hϝw Af2AU2%!X\ FA(/讹!#k.h~t[lp6k7m-`iJ8z/^#Z*Rup38/,9!V`̪&u jTtO͠j9.bif2ʿd{[ȯÜm ,h&e}]^s[YZK{L>k0.1:#XŅbW{}3ݰL`u `?(eZ7`hsr}Ȕ2f Fq(0(R!u:LV}Eζ> ‚ooBY?s|{lj-atm}g%9_ڪ'7>4u.q@}6wuyPEC}.'T͠{WM[ d6$v*0Ƴ_RI7XnYPC{]c<ش3jMwN~lq7I>o:θ^XS=8b]6H߃%aFUya7_x- -8o{@ q[K W\qy/oon6gv}G*|[~$"Å,i^(lYV0- V_{AzhN|O(q9W~9葬6&|?*,(,5ǚyPQiG/z :_;|lLP5ha_W.W|f0k1ђ\0iҔ&0ؐHy`0/ys( ` /~E4ORw]4m&H@ :PV3ъJ5 T;shQ"KbC3r5qb>8,VznLyjPji@kXEGO3g,X3@su7i*}qefM_hvqG MOq8k~Hos<zgr,A絊UDa3\cs>O0Mޟ-%L5: XDSou(<5~>.Gƕ8K{ͱ[s ]}o۶*F;"/v`V݆!(*ةM{)61%JzZ[D<$>0##BpCL&?8>Ql95L;rC EtVv&DJ]= |K&Vo͇ܟRfl&(UpS UD8,`4"fP ̼dm (m{_lz,đ;S4IʋD[xb:—\7ؖW" eA z(^gQ;t808ϿF毹jA/jA/\-h QILZ˘`*-K[8AiU$Ja?mW_8/g8/~DBD]g7L̍֌@,o7Nyx n6rJ9ionsHG:Oh߿-Y΢/ɡyES997J)5 :PX.vLU ;.-J/5lfqakya惓"|r;1 nƯ kƯ 4'\QB%N0h BDAD&l)1yC :6^mȠy((9B_[gmpl7zq}X 01=*J@isVSg/Nsgqarb9V^[ fЈB,RsN,XF.CP ̆/9P0sJdrer.6Zɪ9"[5!p"A{i.nYל[]ޒԒL(\ҭ<^*c7!kX3X/Hi2K$uZ۷G6-uvw,̛2; U(GxT1 ;Kt DRz۶I`N]G}L!oqo+p2;?pnyfzWf!^oui1P>di1朝FfA0=^l 7V|1fr+[55FAYzf{W8MAbWahL2ۯyZHREbI,$Mީ07Kc=d%> FKm,9H>?l}$l#A[_%>XezbGv 2| |5-^qaWgӰOf~%>N;fCNz:s!4*;.v;79]9C>`'<VqE239$C"l9T 8!;鉝{8!}:sH!4*u*~8Ǭ~8Ъ ZuaC.lhՅZF`(C(:UЖ`WQ$e´ 0$! CJ@#qB mܟ;~Y÷nm6네]2"!~RָK]=w`*&?DxDz+h_xh#Mc,~[ ⸄+i}hjދ`eøON{;9Hym(ͽo/s;4J?L08"spv*Mۮ+aH@_@Aհ?R1ttA_ix]΁RIu'g<2r֟ J03ޜ mN%T j,/j|n5CVcUxBtTc-1AK@KtdBdwa,46"qYmgZJ*a*S3[C`e5/Ǫ@ rAExL&8oys^ֽ;b)zsrK7S7{UxcoWWk ]Ja\2"x3)ds`"TOԽ|mM;tԴ5"vumg*rƪpӈKJ2q#4"Y?7EEYs!qYBsRl3 @˗s d5"^.֖y;5ٜl E"Tfw3_)a2{kIJhH O RP4;5 <˙ YDDIF̈́F!xno20] !JGؙ|؜PvzP3P|.'g'k-/7 1(w1nꤔ GDD[:a:x@7+ڎ|. ?={lwʪ "F$£'DדXUĉY\zi)7vpꐔȯ=` Od)4R%R% *fRrӉz5m V&A,Aip*&2wd2)2AR!F'_cA}RA(jCFҖ"&xBR*)ȇO"bL8_50%C`XȈhB" hFXS6+#oNA4δi+sOPJ!#@$p l"I̡#4Ϩ#X0L 5u5,0lUSW@]U5P=rI KFi3&*yE}+USGW ̝y'B]+O/{(`$6SLug6j\C}'+zOVaسBhsάTtnʼC<nlp`Uyx*?fJd_z@kYoIU"O/dhL:1<֪Ukat@ k WoG8Ő@h/lk$O%ϸ;a*ևiZEfPjewl8**Pڂ0?`}Wf_v+\ε|auGxeιQ=#7kYbj í[ {4@9*#1FwjR KT5?i\=dh'Akw(1kPOFh7\; ӡlmic;n!N\Ȳ[3ܻL]-Sxݼ-u<`H(?0.[9іMZāP+};S(lUlbkMR]@[f.e7JR[-$R*Qb>H(Kӿph5&/~58sepXx3BBjŵ^2_Y/Nkpw`NЬDI( $ɾrH)#ǍFGCta'w.m0%h$M>C;)<}ɒq-4  p[tCJ6C$jhTWU^3Je(eB5 vH:g TWwӿU^gK}6 ,kfCcѢY% ͖DdzisbB7 Yac @('LFBc@% ` }Jqȸ2Mߋ[&i:5&/zn%"H:aEQk·D=Q9"8ӕG[8 -:|&‡߅X p؃?WP֡{PCBwy+0(F@j@Veyww܍}K‚ÎWM}-lZZ'zTtw3HDȣ&1jFI%i+ѐ$\QTi eV-b8@mNuVA@qo-#d E? L-`t,0ǰavhtB F}tݞ)QVjCsyqIZjdK++ٯqXbq5* FbF ay%;p|6i*gM-JeD=<9ܨގMhNGOzOIfS=UHwә+ yiif^|2kIMYorڞRmzj𸢤[ɂ0G3 F=oY#>BX92ӧxm:vd0:KU/R]jtҋƽqdscxϓxdk/_>dy~^_y'j(]]{E 44K7F {C鏦ux0Odf/.P>9}6zo'K]LKv06Rsz0[?Sw0\F /KgsNuR|`6qolrOۭ}lQy򘞭GIk^e8}uO/o\> _}1v2׋-v0y"s3՝^`qT?-"T]^uwC|7=}e3fY(:<3͞gpϿO?*!̆nmlhgIy0|a=zsR*+{l9Sc!Q(0"nlCR JHe#B0,DdbRg\~nW svd'˕XXs6mK[-MO2ݾ6@G̶|ڛܘo~WȈoB *Ӝߜv0.77'J3&9UN}SvOBtVcA߲~Yֻ-*ty󦒜aS|}")׋ByRάP#Mofp0@!\Nη.0`bDLa#,df1&0Čc@c{Oc]]҆~ 7ں;K+!w뾏WL1la8E-m|蠳xQoOWJ4#z)u+?J"M$\(v |:䁀u`x($tS!tQjhHD75pTiO(Qs1&l#me˖Dr[SCAa؇P&Psְz{u,FD"2[7:aςݥ+_Vե#I-.ݣ}4y.(qCT nA .dƨ)WS{exܐ3RȍW.1hy:x(?][s7+,=f$/҃RǛ WD2Ә!Mjf(JfK F4ބ.,~FM W$K4g2…Xd!S;2xTMp&*5o(Ċzէ,s[H Ľ4<{ X<(zl;Jm͹<@bb\⽟g#}WQY3 HKZ7|۰K+=hĴDs(3hQ$@0=L!cԩL331+<URUꬢ`Bsc'3ʮc4,}H,p+bSM:$GPF;-~<\oEDMg< @+?7#Br|WU-UHCV6/1a26.d_)VIIRݛܑe ::2xd{ʱJ=-K2ٛ,;+˜i#=j)sdp~6k\gr̾=>m٤ Ex*HU(@"pX㢕n -!DjϗfhLP+ SWx7ŻQ~Ov^M/?0BK;4L٘#lxq}=|cg QHW^t\NM/6ANbof0*?s0}Z>;Q;,f YIi 1Dbb(rZ`G޲0ʕl,eUpňxV,C*L.$B%ywOђ< Kk$ђjI+qnԐo1%#hC3C"Ϙ`%8%2KwqB%f${JILɭ³% 7 UㆡS$)RnK`@6n*}Xڔbg1_kk78cbӣutt;XGbmw8a86(֛2#4xZD΍d X;MX6]-Mxdyy87aVJhͅw -y%`K<ޠw%z͖&]̂"c.TPV tyْ1,dd-YE^ven8r59{3Lg0 $08&܀^=KkfrnNs{F|tbE{) B*MabUfT$R/D|aҷ"8>R)&> pDԌyT*#5Ri(ω2h)YNHԟź*rp)5t\H)y*8;˩GQG?NQ=G}Qw*XǙGgYvy_;0owZҭeb~Y&C}z᝖m!,o">LGu/]\'t #&Yc2}YJѫ^YjCYGWsnPW`*(­zoЦ@:CĺF. X0u`1Ot~_D)E9yi 2{$G58ſQ4&C[ELܺ/Zhюei!RKr}:8=X l\]0'REL2`BqcHʉш BƘX0C6bN>psaQE+9ҟqp,A/,@7Ε- %cg&3]ob+j;Ph)}6I*bc|4 DuicӼi\дMPR.<0Vrڞvq7yej,xz>ڴA30)6`0i$GF ҒjKey=7JkaHwE܎BIwfVӹ+t5Rq])[]`{fbo$`<8"1!d(r #$&9SƢAwy)C j\K# ]ݴ;CY*_4 FN1OңY;wgc(azgGb=O+PdI!r&h!C eY ʃ<0K@@+ב 0^Uk*֫xv`|:Bj=:"D2(0cf0zi'bA8!JR) @TZmP@+A^%C-8,%+O,:u13 9F ap}l@8s:Qau"0yz82gf:dɇEFќ ED/?ي(U6uHdr2FX'Y`60,tU `C rFcaF`(: .Wd];r)M%1kB;F0g! GPQ&h@"!fh케BMuy}J޳c-My@c-ϲZ3y42 W:zej1xK99r}3:и A :κvNDu>x2#] ~~NMeB_;fdA=.]q;FQ~6MiO~ZNd4oOa ^Y(;g+$$SLvy5Sy+ـCנc%ͮ5ڡ-.l^zypNbm6by>nri9S:.pOVVuhhL<\܄|xz{fxKߗu8MZLs˙27 ,}`⨹2X%:(*}ίkRs瞕˶(tuГb)\7 F7))WԪXJ~*z&ᗯ3M_O:1 M(HWtcgUa^wZiUr:'ufxc@ǻ}h|[qА6>6 y!%AՊyCi\4RV zJ@]wu=ޭ$[qr,Z, "k5Q^ZCRź>Dh%f$f=D]FtF4.\I훜ppD3ȗF1( b Y *A~l6qUt:W<4-iuV񏾪nT"p;&KWOմ|È폱*=lK떤I:t_iu,ӟ|wSP/lxp};Q פaVŵj;6aC N;ﮧ~N6S(YBE90 o5V%NՏ#6۩*>?O{$H!D^j3*>@4H]푪{i)n&"[\z,nn]ޢ0 }h${+K>-.Ί~7LXo>|q,Jx!(h)ACpNCp1 R\Xm O KXh=BD1t8Yd4pUHDJtOZGHW߻@+PK00O;f}aFk_&uÄcd )] IL襕f9^ڢcTi>SqQtr[gYdf#yY2| YNS{_x.-3L>wpdR]Ne\IQJ@VZ&2P b(}+\d.yz'z!ݢ@X# wLQ${cք;A@3e5}0+{7~n˅`Ua/os\~jvi٧KK< % dE`9JE{nxdF1fFm TX(]G9>{Pqyj Av!(h!l2 !e$L* (wH_eqydXf7Hܗ,n>l_y@ 3:Ӊ{ֵs6,̀tݮf\PnHh>ɆJ+ǻ[ b)zrhdkNΒ|L0R)0Lzl50L 8F9BJ*R#6Mʻum9#`3*iN>ȩ(ծ_<>؅?&󯣁9?<޹dԮlO<7}ga?c1HE|YL:T̮\\$U-eadq 7p=9(igY:%q-{;~֚7> ӎv:c,.|7 W5JH$/pc?HC )gX**p+ Xbl0;@88F(Hܧ%Sj ʌQLkowa̫895cdԗ}}W-tF MFw珯:@b qZb:4qV^f~sȫpsyU>eDUΏ].43Wu 8{v:'$w7dpf@4r Q0 9 :Jq{DH ebSBp90\ӹ=d@Q葶u9YLh<O!;42}3:p4!r{y8v6E|L j%5τs7ul^혵84Jܜ@WZ""_wV`&@%"n RR+:{x)hs=5juug6(p6xJnxziOR{'I`\Lcx cйk=zSwI6U@+fKnV[sN0T!>D<7hWa>{lU_IM _Z{bM748e;`% siӒXv\j.2^_sog;*8]+Z٨Y<Bj߼8<( 5SNKO6/&l/Ioҕ MlHS++C[0CB,MYA>~OU`lʦ.8k*Mћ5GBq?)x@0&W+jxm[u q;7pC-n78KfWE]vH5(Jn"2c)ݦzl ]jAjAivON[T;vzL04hЛ3tB(aС))5)^K<h32S,޺ I]dzM=09V2F&~L`D+̔7m*qt1"ӥ~*{ShJ :!ZP::Η28zߓM=G~H4W/zJ2b&˱DwK|ZOyC>#= NjnWRXQ؛mF'ԔƢ|&\v]4~h%Ö W⪿<[8Y"#"V%nf|o 7 Q;@ ȏ'afCvRl1yo`Z4L?ߦ;xb}M h}&TW(ݥ$W%#YϽ;ا'S&ڄ w/?&w1l"_`}=(T1.(:;?K֬,$p]ߧ4KHz-@J] zMpiπ19giסt2s1j@ߐ,e#0{ؗE֗@[QQB{sHAr)ԉg"RB} ,I;g}JiL6XB{c`dSCnlovFn"?B(k8:AVL̿xhC P~[D@S"_uV$"17sF|WQ㯢_55 nAJAEW QAMEx:͂PD-M@).0w,TVuTtf1ղµU/7\ȡa_R>y! ?%~YR_O Wژ{<誊uL+TRT4+ D0c[V+[ےSTJg=2G{DP*$@^+PEw?5  iiߢ <'_BqhnFOԨ2Ο{N}xe\W1 E5q/sLy=3{u:=Aտ.vhN$UFu%zUfuX@G}ܒoj|r|C9B4``c ENO:5:]mr~ogb:Jg? }1ZH-Uvv!pV)ޏ`b?o<eDŽI ;Ⱦ]_9GEKзUuvǼ}[u"?an'xZ8/5¤`^[) }Ԛe p9+gA NN**AyY.HT\iKC قok2/L؂> J'U]譋-!PUʷ)Ƿd[lw;xkf02LMb1^1m4쩽αeKjR'}-uVCLtZI E9\3U>6dR=#n^،r9WSjp]{V ᄱl 8xݿ=РKoѱ1/|] ~b#;jJV46L~a ʔHIRCU9JVb\E3 4 %\XmFM~qkn,fAc՜<"ʏ_ɏow`$Tw)RffoA;'LeXiJ'e*jUIlrt) dac!,rcqQ=?z2ʙ5 {i BNpXT<xVT¡PdTn W%xυx  g%Ac᯽6!@ ] $-3C7vP3"WwaI;hs]5z"e]l`q\D 4rQJ5TJy`x7`Օט0`L{+&CvI4D@B 8NK J9(c)RJcsL}I߽?ߎo\F 2/`|Jl`Wۣr0ş-՞S6bFz#%{d䎱/w??½d65.T zZ]Ւ(>Rq|/z=,7D>ґ|*SR}z˺& [.RT'w6턾T[6֭ U4GD:֭ tB˃u;aé=u&4պu!nH)1֊1VG ŭ"RHN)f0[``)_RI.Qm}pm:9Lʂ[6,` gtJezhsg$ )В.\)ŹaF!c Jy,f4Rf5k'G\`:=f\`d&owS }@&#tC&XZ%7E5WӸȼH }  [ig(9T&PVF"?=-FR1>z .r/SDB7_ƔE4Pl0K hqW$hf!c&%. ڲ^|5ow0n4`Jh4~L4*[ib-pDTQ0:A_R5,Ds(KEԗ* "V%t`eJx;ݚXsO93dC#T0y-zICr⠤QŖvAX{cjK Nz@Kg?wTZre=%E) -DPFt7)֜&@j_c .>wA2 p&ǀRG&o@Q`)12Qtl{{y8v6-p?-0d`ZLo/H<[5)tg'U97fxLpqrb0oԟe 6Pr!^X|) ~h٠6]Wo→AG2^C͌fM\j pܓ.[N_m7YA(ݵj7ژ0ՆOs~8"!et0>Z\sOfi- 9VUR3Y\PhZ!!80(A7' }};Fmw@hP_dtlqm~7i2he[ LD'w1!q0Fm{r4 ̉-](][o+_=U8٧=HMZ~3#ѐ} ͺXU 1{q5=֋@pf Ħb\5bWl]9dZXi(OGTG)cN3IݙkFFpBXj@Z2Ǡ@u,:JjLlqrvjvI*_sS=N'I Ne8)L%X=C]+>Q|);{ ^hIx>JݼK,bX61vryWwj#O|Вaɝ $jC NΗ'rԼC;ZIA\[.r"ϑ͆py$q y4@zeXӁv7 'e [&QƖlN\I_Z̐C3 -?GeAz'uGcj`[F\(d v"å2 1E$>NمPhA{ÅM(I%Ӝr8wEz٘{3Ѽyz5Saif0~mpQ[ք A(<Hۚgy+K4̙.5aUy띊RnMwգ=,Z^<*{vNWw6V7kkhǸ= !%8M^/V!mΞ^y{7L1]ժWDNeXw""H7E{X1xMR Ә̿e%)ɢ$@2H[Mzo<i0lMEE/N?":i F+N\3v,FY/XK- ifњHlKLyܦ_[)(*(gi"P| K3𞷻Ik#Y hP$>Q iS1)PR綍9oso⭘r4dm]4gC[w,gb4Ck԰yN0f`<|<`:OTv̾N* H1*O֚~(yNgtg6rVLn]?Iě]ùe *Z"Q05tjw_J5<>ܤ,3p`J3ΞCfh L45}ڵM'cyk҈GB&Y)>+D#rA6X|ju`f{`iLYp`LYŪf@E7 ~*m4"_WO<*"@X)ƅnHԝRSF=V\-l!]߀R:(3u[4_<~[?_N//H2يX՗oZ[Upt(}2_BE-05/pS'(slh/9h]=}RreGc*PX!N/7yl~YsxgtƓ}8Zlxhr=S2[=a  EQ?srG pGe¡&j,c\QwL &KϿ\@m|U{qL弣L='„Effq:K2fTQ#k3[d(rxsf45kw\g]ːO餝{7Z&Rzx_s&$Bn1vHCN'c!IZZI$ 9D+PMg!OO.˘> V d!4t> 93UO&O}ɕ28F(VL b{f_SUk/{ HS{7t_Ŭ; nX~zdb,3~|pҘf{5aVOh5 c _}rvgݸrRy:VfmML pH#\?X.O?^M˖ )anzs?tt&'bX)\WR, +v\@X,nbb3܂JB)'a[$Շ|,yt̫/iv~勷;:VO:2U6,uL[zx8(Y/H9PҷΛh鏪OPwa{,]nU]Rei}TK~UeP||Fgc`ѝ@+ĶL߬dSe?xvY~45a{ VZ&^Glrp,XAOYVZKKs0U=iOoͺ0qn4_,jcU[0G1:piYbˇ2Kl/{M28^o0%KR`3KD. Nф>KntKB@/o^# 3stϬ@Z[ 7v}L1 ,ih!Y!k^|'E>7^F#JT%nk ? (KSh%ݖn#ۦez gG`iUF>[ N0. cI ,8YI( @ "aj<-Xx){HlllVAc_`Z<4c1 [rfb= 6헫OOzIsZyn>yWw~ğY7~Ճx!:52Yπ${h5V—!~^#.3FTP:TŐ+WZ~Z}p{{w/WozKO?)7+9O诓z$^/o6/ggJ| c߬nΙ[S\nӗ z>-pV(>9j` N~s7pi; ';gףncRw+F]O=[+ؿe6?>;Fpim%X,1LJ+aȺkZselT~llT]7Wg·שl:yyuGs.cˏLyGdoNihZxE ,ڰdIy [ənTr|Sh/W?ڒk]}VWf+'XpP^EpZbyom+,ݮd܇jStoj]}}J k#KdSAɲE D;[ {'_6 ]fq0Lܜ`XiF0iӮ9V]*b%ҲmUoAwܿFT!'Yj *!_"q!,_(B`4'W97IIx ]8ߔ3 lO}}:;]ʌ]Gp+[= `zW(Gk\))d[v)-_!.^({_agSܚya:p99ڧO"8ӫ&Pm؛_url_yubu/˞f;MYc8Ԋi*{߼nKdMk4V QNŧZ !I lz- `9mBBy- _u@+Č%R)ic,8N9ŹX0JXB՗Z=%6~3b42@`8 }+b '$hػq,[n)/lo0mݦ~aZ|z.LR rj9<̟gp 05Oɭ䁈lِ"2Mi_d h uIP'Y_Se4eRo!ܦ˼9ݩ[O 0@RǵE1&\_B@886sˌ"]m@k54A0}9P:Fd 1ǒQ+cHR\-fJqb8I`lNc民&v*1F9j|Cz_"?*6N%u'ĘQN}6i1Jq9&2NQrh,d lO>I:I"NyTT-Աؓ")Ȅؔ5oZ±C:D @:;clNe1vUft |=R[G`RVÈ-D͚g- {_ <bf v1w>WaLeƗ# }֌~nMjW%Lyަ/{]&Y/\(@9եmbũNԡ۫[o͚}anǚ%[Ț%o#Tm\ =awԠ*l*A0oTծкMWAՉ ZXS'EjZIw Gy8Ҍ앧_W7rVE3R)_`ÄX_oCn[H14:.Ӕd#-A`a,8bѭjF5 ҂B(s+ьj B@ \H+8US .tnTCZQ %~`ͣa|ur()!f XbY_.K~ޛM1p} *L<]NU#SJ(X{֢&ʝ=Rk8\̣c^  fW'yVFhTJqS nQ_GWo )ĩV8ӑs26N m)7$ 3 !LeHJf79bu <\-ƫ6xA .Hact=D{:B#I)Wǭ!RńIp K}!)a1B\ &TIMRCiNfaYCu,@*d $EHS9#2&,b;!3IF[GϚaM>n^2EfYA99Ԧ{+ VY 3(ݝ5+9kjL`%O(f1u:h JM>1u R+,Iu /A)a.5Cj$%F;y6Ua'+~DRRHb:-""1&R1u*$q8H+RV v d8DT!C*`i\FkQA%wohcĤ$HYw*UƺMg)PռuŧqŕT X RtZ}H\f^¸8N}{?'&?yJ6P$pP"DIUZH+ * ~?RBI9nq1Bp4$1"BYx/hqp}w'jQ3W@+QAȑǖ=k۹ ]ygQ Vۮ< $qq*Pqq*`i, >~RZ?92 >~!0Ys)}P.fO^ZLtr}QR!*$;z9l!D TfHq*  CDKNHB\jMB^D'`6(U$`xC7?H)2U̜t|Mo!@xfovOqUxO,|oNjayܧI!#lq0~gl>>#'_AGI<~w뙽.b5狩74}^;,B%t+@ꑐ\DdJX.L].D~8wd1\NPm/"d:8[̝ιݟ:ڄ΃;H ɯa;A9h#E(OQQ Δc3ʍ9]vhw'f5;QhhJgjk7G:Fw 0-: &CBr bYCMFXP Ǩ_lu|q}($+ *mba:ju A uǔY`BZ:$+ ǍByPFtBǨb.R~6LhYV|"$SniquS7脎Qź]rjf݂ -kꐐ\DwdJYe*k,Y wx n?ПMk_:_Eff31??3!y \7qc;޳Z.Vc*6Z8`(3in2k9:$[A"Udydf^@ʖt~Xd,d<ʠg>(`: ݤO ,"]6NwijYc;Di]W&I*gE ;لOl7YDL]}3cњގߜlm02Ɛ FpB&˨U]0PrE80ԁ@ Oe쥍HE>G@SwUcT'a7HyzOúRL;Ԍ>_BnO\D^P]=ʩtU?Y.n/Z_,ƕj*]R 0҂ptr \:1P@ E{A%C],SI/qlW1H!6 փGBH{s`F˃9R/rZ2Ug2YY"+W3ʺG{Y+粖-J=Xuu@'S{3CymA 0eXQMX?pv5X|И-xo5֟= o}z{5^ټlŶ flX 1L ]$d(N)aIQ9(!8cYMc5~!ox]dCo1IEYҪnQ6]x F>5YF'qu-O_!<^_aTR5N贜zS>l9USZ*)[Nvq)^f)9TbϑN06^Lٺ||& 1~xqjyY} yxeÆAs{3|uE/"U\a%d1`l-Hv!LKp,uXZ!qm[tYyDG6&FK:Tv.8iva+ FgX h*8>v5@v{ڕ [=퓃nϿ]g+FE]>Kr㒠KnC0d&S?{Ȏ_!8ء> %"0~w/^=}cD2I}W!))3H 麺*_PAPҥtbA^SV$RE)z1111ny?o뼈Qi:?E2?u~W^+l/afR^s^s^5{j@=K0KR_Fol/ ra#H@q8\Pm9D!t<_K;$oDvf@@a#p@BJC! X*+l`uF *Jߔ &S,[>å#9>IS'ލ{j*~怾`7Z]-Vj׸q5&4+6ٴg Cp1Flsp^ 3>Ch kER'5&Mt\xdY4`7B|q[Nhuh{;beVREҀ=H.qq+ysT;L=s,d鸰Ϙ, *)xFy&>g |<,|#?%sfֲ?DJJd=/hqFDn`0" !R Hb"惆y` WsKΙo yM~i_V-nkْ-ҥdG9}w3(`qMrimiAb2Z,^0h,h[9V ew^"sPI z\(sPQAAÝ+O9K;m䷧b͝5LP)ϩqA D;ă"SD9 2͵}YtR>9;IS'g{APh˲SaX׌|3bL9Bs FqD]1B~5B6B]O HA51R>%&h=1Pq8^}t ¾93sj5(2 QD@sY%'70ʥE>& Vi J.:`J!;5SVkCC-z\sc!pBH) gD!>XrP*S(gBJbF}yNS8ab\At Ϊ[Vy"8XWlH>ƥ/x1ҀuGa*c#H/!"xl y03\|x+4MZ]㸵~BbB#"1ps_'/>~~QFWCO+BO\Ŧ'f=aw<_N Ǜ6Q kxCeŸ@3Qcs3YȘE:y@8-#*;zn)3nٽ{yU (mz7ah(ڀ(\#4]Qw-$#k{{]wޕmY:@3 (Y,2(SW#KꬕU 5Wb Qo\3Ⱦ ɴ$\iw A܌M4Q`%XrNEF#_#m8}#7#%f9WňoA~ȃN.+3,d=BmDZPKny ;+DU艃 l"\8 *'Jb 6AS(QDZ{AXՐD3z6թ)HƾCfr,$=pbH*`%JlXƆS`ork{+ZQaÿBce fH;mSmEEڊ!mf<^<ï۞`ᔙX?eF{mg"s;g9W.2g{zS\@c9Pr0ڂV4TA.Beeh6ʦ""(s@j-C 8DYcB"ln(9ԭiֳFRa u Ayh13ԀZ~^_z?qA1TԤWm|{jUXUc[OujLSw**v͠n>^lW70˄Dnb'xO)N^y~뜂}sOBV;,\X0(Sʄ,h\( T4C ^L:2N":%|{|>y5T`'w8;8Q!XryN{!:%Zf*w'#y˳z9 ?{vjM&lZ1B4}[Qs!؋})%@sZK68< yHd`t \%Ao9*OWfs_'k*NvPElƊad"ēO>6|0ү k~}([MJ/FRۥ LƤ4)$K!/,LJ-Dw^^F5E !R WbR:Rbg-)mvV<厼D^RjI"^qRj)ZXFdᚣҁ/8*q汓9 F?~2[xq9hcJ]>Ot4Yo`hDt+`*DnӸ z07CWD-{> mFuy;h xi-yι9 #(H!qN4Rgc eX#@<hը B}f-t$obf]u!H؂)jD̐?X;0UN*r sP#or!\H븒\A{"!^83Jgp6Tn]#CG7ak E%+)J]N J%ާ(PRwc*jQˎ-|K="HN-Ztuͷԁ:zAݘա+y$.ʛWv<嘤9Hsh𙍅S~ /~)L)כ ǸI؇9MC`4Zĵq*/aņ4Tټ樚r4 ޝmp6T]I$oWetvs{sQ.2"2|ȴ[=۟\¢g^U9CBVQ5Pj/_x~>`Tot"-:SlLJXh2ͨL\Nk@ Db7zH7`mϾ%@,VBKqs6Oc?/ ʡJH}yzҟ %NE3 NJPŬ/C$4o#RH];+6 HcށaC.JϊBK9Y+}:sSGmFKkյ(82~iɘ/;E`,=bG`мQ/|/Q\8*⵺u6:/9<hp?$/)c.at8WC,)< 0W<z(%`7!>/+ƭ+IdoB8=9<y>TrZeqzuaB=ߵj<oUKM0\ MY~A0i VQN˖6 % /"}k_O $2[ <奭*ގ&G`þ %&:;_Bq^`ǹ(] 0ƘvGZYϝ1(hXCQꑏ4Zc!I#xI1hчOYLPO* AD%5Pߦ橇 %AȂ36~S,<<ۂs9ʅ?wӥe3_H^Ŀ7n]F=,IN%"R%'5B !U*d9(Q$ Y uHr%e"nLj. Rݯ/dۚĊt < 2F)MujSOsd/X|Q"6X2_GJRՙHTʝWT~' '^;܌YK2;ā6&Js p<8Q\6(=~'6&SH 7R"+a"x(LtKڥ5WX4)#UDzb aiii4s |,[L|&Zb`YnzAal<l~';@C1H=qn_$*0De͐ jfkWZ,[-4Ri,pփZ,hӶc'޵s0ASl96`yɇ4y?{۶_Q ڤ_n O^wDK(ژəy  t<v N/}ՅlWvw#9悳WVr6FK,o(}1C|L 9&`2#N!mxl6h#I9Rwgݤ8*rZA,EbAF&\u~\gʘT.,I8-"$-Z`K{oW ܞ, (j(Pb R7[@uô/A \l3nxW/*J[$#İFWdxP줗uD]YuN񏃪z+BF,}z k1 ,(8A ,Y*V0>~wc|]0P;*@`iۨboѶxF;w[WB_F@BmzC} 5Gnmh;&T:57NsPs`>0}U*Iۄ: rZx-o B!_EP$(F4Ѹ7" #Tue΃Q JtZ6ok m\~|}:d0FsnCs(Ӷr!]"7NtA[i+%fGbl f5IM Jݺf B%yNp4!ʤ/;1P!4;[ ‘'!Tqܷ@@rX) ybAPߙ;_-ᛣA9X2nR :jh6yd[ӨrvwY]d\_Ačq:)*Azj`&b7>?Ӄa$ m*[zlx:IsKso3WkX`ts?4WXUgWCV=?O I:e臇HPvxol 4i 8wOfLX 6UG0I"}o0myjD PF?ꊎe L-O=?aGp(2G<,Qgߧ̟DiYh2 Տb{\dZBx?x۶G`]tD@.߉ڇLxŚvp3ٳ2^r'Pߩih\(/ҵR(03T)Ac4ߨ-5TF\Dl:}v69;])oW#L9699~$9`ꔔ.HѮkC3vޒs(Ԏ<}{gPGs==y7#]K I7.賷-ɈYd})(6Έv6\el X)utK5tnՈ1!Ρ{ZalJrFy횎qD7~ngT#Fs# *YoT;$vmjSʻ% @7Ėsh: m}-gD-k,9!:ڢ'k9 GS? ĉԫkj!Z2ۼPluKqt<.ӵ7f=SQg-u .(V G/YU2㷈{ L$Xռҏ;Wg^9^Jwɢ!ǂI@H[O^@@(=7[+@僃(5fDgy“Ev]LyHJ^ /QRMpY~yč. l~!䛼!F_œkj9~+/,he G"{ o8I={e! ##N͕Ap +uP{DZp?ȏɞI- =(Fᷛ_޼*-Pk xmf4< εv()1 Kbԗ!qK ,b(0X GRJt:˲U `}#l"t{ѓ}x1ORrs|X]{J$viqƪ=HtUtغg>  dtk:r.R{(I1%N%pVi$vp qBM3"zIF)M[^6Fbv1+姄~@ Akj֊eŰfcѼi,/ifW. hXwJu ,B)gl>KKx%Pbߛ('. Mط=T >݁/φ1[=5mPدp+udgh<9514?n/#x+<*LJ7NRkum7;䁏u،piTTw1]|;uBl* 44@Ic+PE'_ fׯ 7}{yI921SE "h>[H0FP P FeLl+zeu4;|W$ !ru/&)cR1=|N}{@3¸u'y+݊$ \*չc~:?Lt.[2OF( ]B6֟R%]$pr75 JeS+rѾ PqH(DD*9mGN R}ɬjMv2 DYݮB0%p-`8q7!.Y^gA xTURVz (ǁy H{.ch4 Q0!EIxfjAV7:PbÎ#m%$9ꙟ}:x>&~:*n}ix+L?s|*^^uaڬj)վs}k_s>TVU(U p5wc86VcϯfX^?9 KQdX\ ~L.W7iP/i?zPOZf3[V_vRjCn|KdxA bt'zc'On<۫_"فFz>c^m<6zRR(ڥxSx:oA6wcYޥ.}\=y.:BޙFHǡr m dy7E7V?B{Ytf u5v=#j]ín izS|-1)Iʆ3@tG!\UYo7=[Ϊ;0-O?%[ W?Xd2A%Qa~ /}2 !̙8Paad(lP?.Х&Fp \i_RIIJˊ\3&dRA"%;B68጑ SDL" gښ_Q%'E噜[ZgNR*m-E3nef$xGSI&P" faf)!BǑ}NjttVȄNL!A%gJʬʈe%ΰrB"M 2(NēSЯN]HCe}Ko!W\Ј\E p+cd ("!_LI$r-zQ("[r |Vj[kND?Hq/ nǏq]d:ZjodvI`8 L<ք+ZAv:Uå+}Nj?h`J FNp#Sh$(!sQš(;M!Nb\x=#94!BA鷽mYWxTqM }MŽwcNGԸ->u0>܅IpCcknZk_VA3~p&LOҌ4ùƶgwaǔq.A;zl:*G[Y8͎0y7S;__֖-[~)}P⸇G]`9nϱ->׌7g7yHGv1_O M l>j %f,2Qɉw`i ycA39rHCc"Y!t BHn>wz*yi;Pa7C_ hQ*8Apw1hYV%O(bB1gx])l8}vMn>yv]ޥK֍=NJ֗OHgzV!՞fup'hQqL(qWu$N;7nwfkcV,`kO CxI';~wjw|oJtDYV_ G4u<>DvYǢX(YvuO|pƅ<Udʣ%jJ.iH.98Vjˍ{5H$(@:רQ@\]AXQA| %e%6w7ʜl{#Vgy-7㹦K @oRK>,P_/ %Vq1aTn .DA8N<_&QJ")eq[T".E'2} yx\r ]b$|å#8rB)X"4'ȴD2EHv_<𞣶CEr z=_DųIfJ,z=d!>f]#bbٽ=ElF0^֝ 2֓$M^rq6#)/*@Re:4txM %]t-sNPp;z,4"5VD2gZDLS :ZXJMT||iRB" ~*D @ YN H5ATrC1J& {d:!4zI45aHk?Z1@M;Kd #72-#ɩX~i +:O@/dM~ظiyG %)-m&AN e*`짂 ũ`]550F+I`4ÔRNR%ЖRŌ4LgO]5z1vŭo.&{vmN[;ů~+މʒN$gv#㵿̥y/vjy>-^77̤tbIzyg|6 ^"v-CPx1lݸo:7!1\+M!a:gS.-vR(bK4i MR7a I apQ\?˛aCQ# * h_=A 4JR!;i6}W lt;tsh7v97CkuUT-W?lj~fa连~&|3s1A`v??r#d3x "N27 ^U'd@HI܇ߏhal ْ¥#ϔǔ=B8L9E 5jl%1u4[&zy7`!xoWF"O $0BOgS7c/[p7zNopt aIaiѳjX513v:O_)"6"a\u+imxyv'9N(<:ȁV[2U#b20)4YZ}Ԓa\H:NgkqC~1->y.:pnv񮸬G 0B&j1fci8Y'``)IHQ댴WG'3MV+4rcIZ[0jIO Rb'V*xˁe0*;JG& r{dQY,[TDqGͰy4?bQ,T ,CDdS2-Y2]|LAV2wTQ>$|1*%">k%#*Ź䩣(+A)j$3xlH?a!Q%I*63PLJe'/}yBeF`0z=t6n qMY( s>x}qr3폟ӣsH@ut<RN*nEl0%Gy&AymzA%;K4?T#*\dNbUF:ͤuQ}pPrYGɒ2]%Pj(a׏EC^>^52ޓZ#BA{|ϳrla 6a[2o]NPvq+">UoqҌ@bkyTHK)"ށrGwOg-:mnIj-sz^WIzy1WwkOgK!O8E $^.KXR!ˑRw5jcrdnxQ%0sf(E3 luFAT,ccEȇCqG 7d$U!W`2dJ:f&l .7C_,Q\2˛a ~T ʜYt]] B$YlM҅iMt(R3-22(HŰ}'Q41NVz4byj=3.C(y#&;F3xW W rcRsy'jdu̕^sCY㇗ Ofb @+K{Ϲvmŏө~n*LW3S\0y,?3p_OO~?ccpA_?=xK8Q8\Ÿ#zDT0]聚Sq%KrJ7 G@8OOeꕍGɽ!'T A#nLY':[|I։/j2癘e@|Mp>|cN+m_{(r~޵#"`e~Js{v4-$'Y~Ȗ,df-ţbg,XbznV,Wm]9՗a>)lZqρ$m\*NmD^"-щ:xWSO,|ҬN3f0fąϕj8Y "6>.͸w{k&r!&}d?} jv-(kAw!Gɩjn(`cB\~fT%m2J vtչ #|>#/_xzւF-03^(/yĂv&c] HR$bwbU|[/~q̏ h)Eɡι"n: i+kvRgG$_Z e׾M+$j:c8I"g:#JUΪE؋f]?{acKg0FsD;v^rÄﱜU'68]d9[\pQ\o \g? lJ7~LHLfN2p"ax[*t9N0zФ~6e@txcQFQFQFQFuū\KmkJA9Z %)= *5yY:&cr|c@nl`$:t!X LNRң/0( @]1UƐ1c: *1)wtL, 2^H\7S FBvS>UkK&oS|޶'%%CuǶ }~`ٵzNJun_B\UxZ 1#>JNcŰq{&fAФaH6u<3q aq]Zx[ 0 #(l%bU2* L7`luNxwp'%v/4( vzvnw85zaqv抴po~w==_`}l'J;X¶Tml'3cFdט0 {t0spQV]r~wBr7 Mm׆>ѯCh͜tSv_DNxCfDf!aJmĚk e26b,pUf"gRM k̶=;8rTNIBPo|D\0ܟcy8K2-pt,Nu,X#";*KC%2 @_E<ځp @ty١ e)s=NyWF󥨣n;Bwf3W:s_o5ȚvTqQ=>|<0+Y. RHy }ncQR¨ 1/|.\*4q++u(Zv@(;+n$H٥q/ 94nv[Ƕ^/Vx>~rSZt{ ){4n$?'btD gE0#l`;/~up;o#wjPr9rAb*{wWfl.C_cD؇)aNQ0LGUEUEUEUU 8`,m f(H7Hc L)Hp QS kj1" @]|K:Tx!Ht밖̶xo֒IYSz49@(,{Jc3E- %2fB؇ɟiWQ3u2J6\ _]Ĕ@lvQ=̮.:ͨJI[k#%=7 KIuWBAZrT3rgͧ:T-|c&{}|zn6 }EG޽}zśOVAr}E?~0KP B7@ nl[͍,ɒ?<݅w-@!hnJ% , '}+W\|NE0*2E}0"aTe2 QLx6~g+ŗߨ}MѬKx͙Qs01(B!VT¢S϶L E)8FBiB *(bTj6<=8'Fl?gaN~=n_jm-w᧮2K݁NziXy_0#%[̧a(o _#NadrӴO.=_'^1̿a3DOuW$ofA|JVd\m}2=cJ0k 5n/Aonܗ}f'I=3kP s'qwn}8Yy%]\NzءxcRA9:Z:˨>G5rC%ңs0v 9^*@N>{ۙrh"YXDR5- NKP$;*rٴH"9'S3rb=ԽSM|3|OnxwwO#S#FJVql)İV\+MLI=:8ڰ#/:ccJbӈq=\՚׊ _ s_Q#OsY*< Dh A4Z9i EQ, f41Jsm^;Pqc&Q)Zۜ5%E3n&וgD3X-eUS"V?Eh\ g2[\lNnn,v ~m̓_/bQꃶc,f .M(,SVsF.ZQYE[6i"_^ه.BzUyk?A8$!r=\ n/%(PВ+GNŮnj@o}\>_/~h)Yk&5s(BZ3c|\הnj:#_!(;&ĨzMb6_91|LS"2K/ 6!6!*zoю߭{~۵Az Vw}P5'E:Xykn?'If`NQ+7պ5ɐ%bnE+W ,H<9ᤛ=v#::7?CZ#cJZEjO|qxBKe_٢E0 DtmElI-J\D +VrמT֒8RFy+䈆1I7OqGNٟSVzfvk]{!o*uV4 co0rT/tq6g&1>{//%Ugp]@OY^ #w&>3 Sٓx)Bʎќ]S x-9J,rHCgnjv\S7ij 5@ 1)UZHKN񣮏^o,$(TtF0AkgJ1q4},x,BLR{߸@`IgaPIyb[7Q2xE <̠p ;luɴ403fOnR4ad*+D4#/;mp;A L88NqBDpB(~[E"@F`wX@!8\e:m{ Kiu*R˪O~~s]عM{5dw[I)^bPhRröO"Pge{@X}5P#j!VecHg'\c.a=MKf1nĤ+&BD'o77#%&RS\"=) ~a[Og{$X-*Y:"E V I+ж䒗1D(AYHԻz㣂anx+Qf/ +IaT` - `,(&L4~5 |AmtӨfa/eደX aE)M8F!Ãi, cj2I4^wL{LZ Í$YȏSga htOv ;wOlS6unsf|xx62R5 E>{f&Kd9u#E9ZO^Q T[49ۨe&HЇXORiF󮰪^5ДO>!S]UJ@k]5^yJk75]U Ϝ]Յ|^ L1F62` Ӄ`iBmͲ[_؂[<^өc+ch>Hd/!.IJלo4yi*2,rb~[^|8˼a[缣FHN93|y'y|71a)\)sgyG=$IkthST)IpΕ]i/%-EF $դdmFrIK?p煋.`4S3i8` 71g%z  eO? l^ 1^x[^ <i W9q-8QLpҖ=olT$[k@5M7HwxϑDRX~pi3io-j0 4g?~ڿ`.qljV"JoDBZsgDD1`r+r8IKqRmc "|/9üvu Rm G,XxkBnY3GD`9hT"qF'G^E3إ`(>}4[[ >tysܴxuJ <"IaZRtpSi%pCUW5BZ+h]`R0;d-01DdFz4!E"~U 4YGNQd@ I,@0^76mD?I%H(9EEY0,r=ޢ(*XQ&fyLf>"B:9xd_N wI5罫"Rgf>m7TVIw^-2@_U=U6#}^(ۣY:]\/7֏7ggt3#O;gzfLQw.72HNP}yv;!JmW:>_N%T#J|R$ƽ CIhu!qFzmS8Hܦh<9=(YN:f( `kPklҎ&XBpx]t/5 ےB#P|6:&Q`⣽뉾 `i%ţbT$:'1$G)sߠ3Łh%U-:ɠ.`%j/I$xI & @IWc,!ru}) NHRm=.Dh 7E<31OpW";pNáVPO?jS8 j,QH 5c$ E<:u4LJ49th|RY9!QYp* @um )Ԝ9. JRg_{+Yˬ>_XT?v jr^A|rgu}q9ק/Wѐ'Öی~S)Q)Kv@NܗZgG YK\foˎMUɺ?U{KWᲪ/~;T9}6v(BlRϸ8M_ jI 5W'*PTѹ!]6(P7q۵z7 k}sYrW 1Y,jlh 9bˎ!UyԖ>Yz奨>۳hk7WէWxoUW[WvѫvԽ,RY ~^m.A^kA>-s 򚙎;} 7t ;F>ϧ9oy`:F>$-sP@vq-+⡵Q=qO}-vU 5j~nT/Tϡ:1_NxxrCI@O.238H3jd1X"+?+*_|DqbT;G~WAe$eX.F/> Ǩz sit0ڞPIs "2񠉮"r{ y~"#yf1{,DTW\è5L5>^o0EN09ZȊS >_shR |{AJS >=TC/im2yb 0:8ӍU^ECZz|u\T~hWE9n= iSd՝ SǹY kYl`?ފI%'hTyR rN"3=͢W rl)tiv?X6na*2MHza*D&h $pB%X!!VS6,dQ;܆rkIW٢Dr  H)Z _:j88\(1^#1֑ ] shOY$FGIí*R}^@F^&Ɯ qZ($ZDnS+xBՈ[+Lr9+@[i)޻NvcE =IRcxۍ,#E壤naNx"EF*M҄zZ2SOC2JGFI%Ui<_}R֜V/@(0zfu2tp̀zKfs5^Mwg~PP1WƤHA0=w)rבxp%wK2r;t}ArL:+r ,9 {d/^NT+!j\de.oNƅrRRm%U[ SX<= 7kV$~a'B wR2׫W]G?G#L L}G/?=rA~9;;;U0Zp=;P*Mu`XHuhW-Pf:tsq:PhΔ|:~ W)~WY1:]6k z,srbv~aFve}lDMon57y#}3AsVMMai<.W˗-h|X.^S/2C凟^NGK5Q+Z_M;\/zu*wxdHѼwuћzzO[WfyyvY Cir XT^V-bw(kb9/p%Q!D k{0EBpyp;VYƭ$N&B$Jyӓkm1B {82r-mNHd^Jrc`'¡.0cG=A(s_29}@t.2FNEHխ8:cfEOɱ{hY|H {45Fqspg\CnPXkm/?ETFEAp/S*_hBWp*89(Fn<ɕJn+<ϭOFSK;)xSW";\ >!hyIwkL>7ZPZyo넬ѤBkB jO{*G!v`!Jӂz%-UZgJ+õ+zeSD42C)еݵj$ABDa;$4tYߗ`㙚k*ےWxÍc)3b=~M;ʍ>bzV *jDCa?P㺊߶%wHt_ʁE~uCãutt[hblD-^[c 0)ԏ@ ^gh"@w}H}U3уPS!ÉP3vG~)Jhhe47iK;=Np"l NN 5V@dd0=4Y[i^3 ׹~tZdÈwcY6Zy>Ծv;z?GFɎ_>ѷNs`!*$+FJu*yyx!{$b@ussDZt8!:1T3}P< a^,[T ˶> ps _Ʌhn7iɠWiɵr)9%O :ɏRav8x._)|w~ũBxw࿍G_ft#cncsx I,-ڦk8Cә'X~U˼Cgb-2Ҁ`wg1q՘ Ts׉PGc_*D7;DgpN<*BtFh01vDţBygX[08=*8ugӖ0BĘ_TM @$24i_HF !%nAg\J $@S N|lZ.<Ⲻ20ؼ-}\u!O77ŔT. -u[~׏0lI/{߼}w?T]*sZon'"Y|56x}5_}i=p})9~O'|E=_{K+<[udkk 2W=;R;Z؂Z*HbonCR׊CJ9()jϠS2j6覶y5auz~T2|2FoLMuAxbaE4}2$Qh<< 焁<4(g"VS:d:Ͷrֆ.<["^d ±Q#V}(|1=58mZR3JtZA@ 鵧A(BBwQF[T*PEUG 5@JTZPK_FfWs>y)(G2zŨ$pҳ^9z: =݁;=/wW̠9_69|iE%HdLDdc꫾칿坰dB]ċ;U_{4Ub1y?%szq?3w"wlgA'sJ0 )9NjwrMyu۱/; H/l m eh)ew5wN}`ħsXJа*qPcYt4JlTanhVA]]̣x#>0gn(J6T-rJ/1#Rߚ; F;$1>ۯt6V϶ȤdRT1q ׭ Nsn,Q2 1(ɕB*>] sRdBMa817pj46{ :jYR#Bpέj[.dI*%rȞ0ZKt)Jpn/C0Y,h@`-)N`m-!p!X\ϵ4(SJ\+2ziQ侯bAKO4c%8[.@4 8NF[3]Ȳqs ߭6\f/κZl_ܕQONkj4>,Af0YfpjA*tI, ɟsz7|l,LtM-y&]].hl AjTOy|脇m -!ϙ-ff|`4(~0r3(ocD*me}sGOph63%; H-:}a<#> ހhy! xSm6`:R6 >.%G^%f'C2HYҁOT  KGth A}7d*Y%Qv+H"S尖1;/5)1#$=a侣Ŷ>3@oW8(.SхSQJՃ-`-KoCv;bj\- .=Z\b+jM"-X-BE7 zv-v J}γ6OEjlqE]W\6kd}BVpxwNJ_E.QGZ b1Ar-@X.h#Nu0$Q| /6R5}}.[3ǻ|o7[}粑HZj?YZ |wt̨$ 45{g+B[|5~\!LVUMwF?(M#o znF}{tJq4yAtvuѱTAQo=ƔI[k^f<]8 HܓSP gOgLpry[~YfOn #3Ub!5cZ oX3SBns5x>r)pYV,eq]R4eEF/,8^2biS΂Tt3qz=Z(we#I(]݃f靝]yVi%{$7RAHIRQe}q gi<&+&:>p&;5:0 g)n},N;!=D9'4}p*o"㊊kKF.̓ā(gV9'R~~C=a RZVsUr{Z t" g]ZkMZd(#:aIk@mֱ*uJ$d[[, z/Qvu+^1(8[ni !ͤW|~Y?"ˏ7>w n&]iό~D(Rÿj4C6Rt3CJq08,ǂ8S:caI&"lQp!a~O2XjI~"4AWk/e6YlJN{YptYkT%#s{|0( -ⓒ6Wnj%r=28_į:)I&^87Obbs."ZLMIv#YxrN8uRbi偮}RBbP]w: + IO=Z)Cn˗a,˿3mI&-~[OopD"ZmD\ Z2c'A;&#Qzi{+ߣ_0߯f Zڹb(mm_$ZIoUWOa;I7nL5 4ivSyp ip =CqJsq&e#4xX~Y5\ԕtӆM馍o͏aXt&EjtSz+Kd==WI:;zn*Jp uE5єs.̬e4BbPi?X[ep10DK1cZ5 6/ݷ.xC5U&)iLqyA ] & [tJ^CBW)Yr/|F0`.l &~I2khScomHi)!LɳƈxX:"6 ,pS~/@mR A|3zY!ZM40+=)8Q@ DFHr)RSt3Ej[(uτbKIoB"*oz /`hG~KV:7AS%@nkrAZ_Q%oNk :^K?Mީ5ࠕrҕ b#0CH`":eXN>1TrYc1w0L>BRqIL>I% v'"2 J{v;)؝D6R1ڞkJ)` |I|NWjU-gKI u J &<Ti@i[ƤTow/>S%2F PР$&$c:jGA @d)^{5A\tވOq05~jsXΧnQ+3maM,8@#kR1.I*"WU:o5KG!aĤA{zֆ$SV3Wq`팋ijq*a`C`c%34k1vFeuPk#l{kBnptkHRj#M)8D *2*!n>xU#*b?~$?gNasPPT3l^O6򻶟j:7s.둊-|HdjWf~}eC1yR-+'aZV} Ј .f/ٛ[JUAޞ"gL sbO#X/jJQ*ф#QqC }- 7ZH۱ܒ`su!FD^=CփW.~,C#&}ɩSK E} 㽰и߂KoE)g{5aŚ oc@{ .Q8g4Bl# 5P d}OwR76zcb`٢Vo혒vL3 $ީGѺv_}CXt{3OO_Prӗz)%('LX|H!ٲ$ h{_Tיmd"kVdd9R B;j/G3 P!30^ML)`W*h plbWPlP}:C#!4%A5X.=0:т?7uWeGK mdzÐ t>;G/xhT^ᖁxiˠǙ>^> #A$'oՌNm{G7:x%FgpOI0y}((:Id\ǤIpr_{A" PTу~Aryj!ʡ/m8!/\ǣpkK6=GW'UovGkh`uź_7H(1H 6Py%>nUR3Zp??%05vB#GE`s v;'GN5`B՞E̻t= j}AG!a0(1fGǸ!0uV3p[BqŎ!`g\L@K0rsNF.V1F+CLqZI A!Y83,2"P#j1rL`^"HT2@-΂W.(e fAJ[ww!?a 4qc{g_cy ?%¬~5X ! ,0 Od`M$\C(ZU ݸ m0J3g4k'e &p1kMDODw@  FEdd)Ӓ Zc GƁG Y9ouN^3E+.cTi :-1Aֈ1Pp` KaH-"E\Յ ŪU3 n>Wbx` ``b%H@4s,P:zV!Ac/젺{{ͧky˂w%U!/NcWW(zW+{VesUs9 ⌑BL$uL*X80 ъ1aa Fm>׍Oy鰖bb]HZ{av1C]]̐Зj3ŌEvv1 k3Ze9KCe)/i4Z<W̮l/_UݕWphu ($\uϴXiyֽW랣^O~{9UjsDԥZ0ZT랣^|Fqs &pPޑ/#G"H"C\{"F)~]])jsz ˵k3/.f\]Şz?{Ƒʀ/f_! @668yXB_-HvH9?CJ=dx0`[WWN;v\;5jldfz>dWDs.Rۄb%!ʖFZBJX@J:vDrk-q^8s^ n6!}yStB.*-8HI8:$,QP`֡)L{Ʀ怅`b]JZqꋠF.XsS&] [4sh-$Sͤި!ON(p :F8笹cH-TP͗N@B5igU*=mO^ 1OUu2w|3x@ɻ80ex^E2oE -'1MAj3m'Նv[.Wm,ٛA;k^vl+V*;ߘj?M M~|\;AZ 9;z/gb:\8QM?k~ߜVF]iMZ>_wK+d|C]ٜjQkQ]TٜjXϮ;vnIu}Q?ήf`;Gn{co>yX8-)Zn-hs)mFuvgsvMͩHMmT jTSٌjT'>~BF+J0jrUT7(w#!"]ϤJId] >ͥ/ƓYy 3yf,(Թ].P;Ɯ4p?/}Mg{s@A)kbB4Ң^ r;r>6կ y*S }b bwAcﶻbȔ5һu!O\Et/4z7>x\Q1n+L5ڞwo`jݺА'Y:E$Wf[л䎱w]9TyŖޭ y*SZ}~ջ)-1֊v$iK*vnBnu!O\EstJcߊw#XwAc;dJޚwfһu!O\EstXnѮK41LwߑdĮ0Dm14䉫hN5מU惡ڏ-Ƽ1֋vdcܭc| y*Sa+ލƑJ;ns$ͱ}^߇NM]lqlgZG6je@Z=}O 0v&PM{A^N-nzP.g:)KY({ǘ;Ea3c=ܥ%1Ksz[K`,6SĘ%0ǘ{SKДƬ0'=c]ZB ,=c̝Zx1f%y~1?YT+%zǘ;Ma+c=ܥ%Zc s1wj l ;Ę1Bc]ZFØ1b[<ۢǘ :[qw1RzǘkG;Ę1Hs1wj `13 1cp-1c̨1c?}FcĦ{v1ci}1wk zӾb̘/{ǘ?8zǘ;&o3$Vs1xKC#f0r&(5~EϾ֜&pq4HoE!:d,dFKɔAB_%%q{͉) YT^2`4> #3(-\=Hӟ= h?6[t MՇq2cBxIUz=xbgݸLk|L1PYrvakmYx |ɿr'[y6Y• a҆;n=VEgDFa8d,UX(ZP Q6%31Po5i¨fD08"A`^; $[1+&TSH6EΤD.aHlQX/-^`"8рXNeP ba/shɭZx XT̕1*8ӛS $L`ƕU YC L fr4 ;,X2kR9CND Nx (8͜a/ ql4(4|!Idate!Ch )`hEiIZ )p>-K 4fp, :eJpN`N R&D!GL^J9*u`Y(2*x!,avpa8Ou#%Z\0*ɴ$8H);#-h=VɠWNA 1T8HP*#fjxy-RTӔ/]] ޙ<6xus )]!ϹTC(_ll,7`H_jno;"Cxr[KA6v*H L,sdq˂|\Taa_]&۳P3rĆgɠ l<2Â.T$flކ|dG yU^^,CzǞ(D#AG@X8)c3!"Ebфe=X`.:Sm EKWeK3 c|VG \^'X\]F&j wZ95i?۟xL2 @ɀ "JDf_y>֝oqyϥnLf<V" #u+]\ lq7,{3s/?sɅP,!B?P}Ғz2ds&ROUJК s :Q9H!`[!2J".߭A Ro|?-Nk)@YS)p?Űp5|,C;o빛A4"SBWfz-x1'[Nb,UN#ܱZbH67y]c-$)S`LlMvPePPŁdCK5vdG*ֽv{=d<0HTq'I1Q[E2f/ &ɫEWK,I: se\H+jWǣބs/⨹&GzEKu@ԜK. (b8RygidDVӸ{jiZ)'zm/ Lq:-$WK1Yrd1zѯ^K=ѧc?}jÂAYF4ջdI,[bd;Q=fX A2{}]Lo{S 8e -@t#~|d,Kb>Dk~r-a>3ߖf`Ɲ9#wwXO294,Ǩͮ#GX* p<y21,Ѥ4Ja 3SO,TO@[iOFR4W_]L{&q_Q@ū2ҮdX]=[#=:sraqs˧ZQbCM1P6mCXI֩t q0(6:&!"~j,P6wi!VҎliEߘ+`h]0c|w5y7g|LN0U\^\u\Ug#syw9t}mLђL1ĒI1/̬ǣjZ՟oWu(-BPZ8M&;"ukdDH <K4"(8(?1m0DAHtU (Ƽ@c4i쾕p=xHERw9y[.:T:[8HЖއbX?";/p_>{ﴖ$ ?r{ )F{HwXH>u8яaqKi0M-RhK%%hDd ˽n54ֆy**z\ `,밀Qzjt:f TChH[lk'ȂP1yd O正S# V $#ѡuXP;Nee-9) ;S "bN"2l)ܜ+gB4[OQX\aPY= [ϓqIW`nK3d0Db-9!^`VxQ FhϞ%Xt4אn:,3 ,/uqFKRu$#8Ӟ[ o.G7R^֫AH0J*~\s۝k JV7(O‹LG:m1Y]^j f~ +j996?k9.N7?s4D> ߴkу56.oz.#%ްp@~[^ svo6 $/NEԳ %}jB|7R?bѿ8s?eW>@tc.]n8]A}fN4m Y6/)lÕ辍3[,KV>Ivebl¬TwJ@qhXURR9t:I@jG謽H3}tj3~X=9>0ZѸwC0jUU >x*384Z%Ƶd+zp.P&`2ŊDWT)"|g-,R6%l֒( 0P$p}KHhŕL_}iߘ8}ͪr 0kYO-1{JRf۲Vl,2;b}ʒTUk[,H .kվk2}Nz"-ӼQk= YPz6~Y,2P?eaa6߮N$0AbZsqV"B_O.6~4Q͐"!^UЖRt9e;|KBt}2gLfF+!@\pГ\/i_1vg$\NxLF'=˜`#*P"C=oNe}n4@ :řR%bq<6%L6/Xx(#V):,\pMB{}9#TEq ҒehWl/'(J83nweP LLaFAՍʻVa׳ifl\#)d]$7+yq==,,8zc?Nݯt~gaJ -PB3Ȼ4WBiHW &hvQe 0e p \^L^_^ศwo&/y 8'eF(7RZ#sC¢HJ* ?99A0x7BN%rk8CsN!ORÃG%QE*CR.D@k4D4.=4ѨJ Ehg4A, Ð7cTRtMeeFDTDs4YT!]6Ky0a.FZ28G2arf\ )`n%z|Zb;0ed២R>%Q9>޷8cBL-o}x@9{0{G_eGS+ z:fȾ c/pGi=jO+1rA_=dC b4`t'phvϔУ^H-=α/M/aei7sH.;*E*%Pk $]8]߄%tJ5BH0#*'*7-r>SO-ϡn.$q9OO6qN3esIN N* b Ea"Hrȶʀ2$MLfoQћJd!GpUDr 1EOfam.wЋ(cN!3NzǴ\FSIG*wL¯ )wa~Ux_Q5r !N.1hxuvm $t9aHlDe,u. =ɝVN-TL!;E*5WUuR׫NcۇZLoTqN5TqN5SMo%M6Qb"qC 9Jl(X ,r_ ?D(M&ZvVom7?œVn}p#1u>E$ghDxKoMk>EQwH#<L~1; dMLԉ(4ADhmƫt+2z_/c3  mA+Drcze7`J0i&;iR (4TR <8l?sPa@\kQDJT AQV!b wzPiQ+CG1hh$]^,|wuz<ˇU7/ =mq=Ӄa-2!qFe7_ |4sC`oʞ~p<@>h䏾t*_ [֡atH mPH[6 TZdYI;5E6='sbTځRhkG`[4ޖ}N>]ò0gG]Bl~xX6-3{Aſ윈}7cD27GT>)2Aת]kr#/ Ag `=rx Xd2o}d=ZJL`&HGF%G RBϔSS0Fw ͣP:oS"(;[aRBd}}>bq8%L` ZQ)1',*ʹW0VN" E>52[:ʼnC#N |ւBX@5h,wrOhιиLi\4 4j N,ϋZ fDLvt#əFl8jdP(#t^NAeM((تCR2Erk9:(Ccͬ:kj.6Kf̷!ws?,7qi@@8-Է,u="2\1ɛpBD-#=s P[4pd\CnqZ Cs#nw)i~hq TCrA0hVA< Eoσ 1;X=aH1XF`ʕՄcr'{tC#jwi8twgA˞GEԓ "uUH2p@n$L&sDu"I~7.U;2ˏ߿DsLbʼnlvH)C7ݢg[TY?\/ 8}^Oo$8HIN,.gU 5bEk%z|ǎGDALzr ay'@){߱r֢k,޹ 3-%8?DR1p/hB t26@~GMkɐv" r jouPz5pu;tDXA@ ;+~6-7>DIS1QdZh׿hrAIo(}vSKeZ R|q1k.j%`]&kw>Mvw8?q藄y& e_ntdw}#|ω/r\oW{lC-]O%Z]Q +RTvoC3&H~&&GW(e[w]Xl= [t~%Bt{ D?8f0^nZ:"9"b#Ot y}'<:IO#s|=+ie@IdbѰqۏ__ [ "|Wx|YJ[hiLz¢4Є_QNWBOyZq9@S#>Aݵ$h4: '˝H@ J;k_Cmx(S'mxy!R0˼RD\R3\P*`"Sm~Il }r5W(.6s__\t̨ӋVh_Be_6BPdw_/ʝ"dQ.9"4x 8=rZoa{|UP=4$Z@p4)~ X~{p*rgMuAFoDD;O| aNLB|BTND垐5{chU# Xä %3nr`Jx&&pc 罎"J&&m)`A<j58b,W'$p 5VЏ)Pm`1gZ۸җhT)۹&\ndK`0֔ȐgkD 13-)ǢF~ht2LFm}I~\ȕj:,SKAU\7Ӊ]`ܯoϿ 9;foһ%NSlrqY?P3[;ƂUxy9v#l~6yzA]]-B:I.|d:vjcX-vɚJ$R,US}!5R9![x0HB*z_PP|u_W5B[I-]>t\?Amu\/*f0ߎ+7trS;XA(kX67rtk.C_jz3n`e(c> O_FxW4nB+fK6"[ :q39HN;jH{eG~$;5h?.$ >JSoVM*Diԉ9 u-:| ԇ!]i{Ż(#<ǻk&D7.]{rU{55=+]] /71DBń Q{yG ֟{gBU`0 !*ZċQ0jV!~s"{N5Okqk\qPBZH47/Fkl"%26$=w&:"=J L̑o޽;\n﫩qag; ~J* nfɺyJvp =fb;f/þlذ!}#^+=ENгwu;ωXV Mx3K)1QJ靉j OM,ȓ(xUB"S96UBvzt^86"ᔮb|6K|; %pE&@=1O?ǚ vhI15gX䊳9ڨp[wQj+1? tpwC8Y^G<|[o$ӓűbuaNs-#, kDxoqQ"SB.- UvBpdQiJJaCݺxOп2u%[7t5X?`Zofim=}DCzT{Z^Rr^CH|t|r ָ@r DT[iXK܀ԖR#_$ĥޒ#M#RIDDY$ \@7BMi↶CbtV$ HF MY"1 [( 1%zck\7&=|[MF+(4(CڅO}N&xi(6N|+N`3!FM{i$[04{ If[I&wY)qi:+ ؏<@/)sk ߀`L$'̇ a,e˕ )He B=@-40Bk|)Oyyy7oRv{A2GU )BJSdVRs[ -F䮢Hs#,4w(íR:`?TH[1;Z)PI87Yuxs΂x0ׅal z`S%7$/>Dx/.=Pԕs͒ޱwf7˒Cج>Ld]K+]K[f]g]I>n:B?giS-BXX]uH ӓ4ixY=ޣe/>g w;B5Y []tRPЛ/?c!Tv6ӽ.vi2oK= pҮ"ETyZ8qոMiџ]oاQ8vLϡ'wgdc'Nihv,P(4̦S1|1(R,MDz#ToޗOĄΕ;O5*Bm+R?RbtMFK4 856;>L@w05U1qzgCjN~n]bޚ8>T.-lANutP. 1vވ?8唡7k,I达cU&IP5`%=_p\:41Jl1* UֳVW`t^XӀ)8gg/TK/ SIc4t4 ׊ ZnGC@uvېR%S3%!!"B]>w+TYEC+WF~uYzeyX ӍՃO!Vk`HoկJ79߆$0{T8sQ؆{y1r XtOFa:8$8B(pK`ZQ9qT"xW+95Umuw<-h5˖rTã q{=·6_1Yq? ZD< ESM]vrq*_E*ݧ3cDaNmrK+rw碼^F *Tt, k(%m5{d3K~'y4Oղj|Ub 姃>~'iz_WLd竞OgEXGz~]{դSv%?- $[}jPbѲ___[xd WFE_\+ÐP~e$ ;pEƤPW?ܫp&u a2\Ǔ!V"JJHAe۪"I@hDbsi ɝ;P$wG<PmzaMT" ))J N:I-81RB!,K"-gFLrҰ/Iعv!le`# &8b6Z))O2:)ټ$a_d[/nݭDI>TJ9g_IIѕ X.,؈kij3D b-[&%-[ullUa "50L խqDezBck$ gnl246(iUeE;t~U5:ɥeZ'fJ"h w&^g(TQQ'=Aq2Խ 0xi!}p w`h$N&Qdxӽp7-4(f 8mUW pNkhg5183±asg '̝Y&Hɼn(i5P0')7$BEW/y'*Jg)G[w>uŚzMJIB_P2j,oh0Wɫ{rꪇY=L@h.֕vN=$c NZ^hC M:v9͵ul".‡p\ 6XJ+KCT+Y)Ҁ2@US`jE@gu~42l.W߮l'hvP,5 &0kxe T!1~Nq  3-+>!|~zOFz~Cw?bJTy%Bm." > nzX2>=m|'M\\.RQx^X=,T>yO4s"8XYq)c37zD.2.j!5OglA7~O_sJ d0wiRcU-VDgTՙ1.:K9*?t]tBIJ3 ծQ^Sosn%8+-J"EOޥKs+XwhݽFk]i rL\gmΰ(H踅5,R FDb9 V*,?XГs1gTc컲vFldQ2,'ˀUEjm;p՜( Ź#c *-Zhe~>sU&3Ɣ,`FPgOabu*P%G~ #=fMR1x&R2DHɳJPWHgIDL[t$;Q!*j?e9&Jg;_Z3qv:Bd s t%Bc]C:#; a0%sUtN BcP5E/1R0[bg$(] $Z VrnA{_ UޢOW:뮃}] |*扰F2*KV'e](}&2JdpF@+lɐ"4w,7#֘!ǒ2ܵ8dn#e9LqBr 9%xߥ_GvrD)C(+,X+%3&1%֎bD=Ym eO蛋݋N90$:Fz\{:|X5ƠM=˧"nzjWJHŗ||=ynԬ!ʕX4 ? wz&o\'@Ak.s:/Ȼ]8bPۃ7QR0_ Fz6vؐ2$7_'sDb0^jLJ{qg|b7b>+z'D2*62\JO1FW.])(N D2$rO҇HV G 36Hey5+n0n%q]N+G/!sWťb[ٜl>l`$*13CĦ& NvU^)9tU}gٻ8n$W:1K~k6&fvgv^fCSj$mknbU5BAJ}_y!QT;D hakPVgf ,f 4TFUfd vc'@{2tn-|zˍ|}AcJvd?giRc:K;e~b> 1ʒh<^N2֨BCP.O.^W0sN o]VUbj'k菓_槧%\3yvKX uިP| 'M1%ܞX+"Ԭr6sJ2*%LkZA^V}A`Mr֫ P1+@tˡf#ظF栤 c:'S3HkXI枠V!N$VaY:. nXo!jXqk#k.YDݸ^C@"Q ҦA|r<=2x~>ՑfGʵsyTt4U=o|JvtW2 ~A!Xra>>rkyIqA 7,rDW*@]$fUt~xbF!whOzp A)~<7T pkY:#t~g$bk8z+-ckT^j*O>@+F߃|)@kX#Y[]V$U̘Rrȣ_T|/cg$g\CsY ߆i#(WLwnZ5%Wu. U}w@4S0S3LeLц2!{qÏ5sŏ^#?Yif]ZykqSNn@)Z( -tܙ7#MZy~Vtcxýqj0[Tm_ʦ0 o5[С2sSːO 4:ōW=~ybSTǘ[7>l@.酋i1V3*ͲACPpO/#_4Pk2@u|Gk;pԣ+5k]A^z{G-^G`VӪ{q+koJBi,h>{=w3{][eFME+T RSv!\6I1Hs楆'-W@!\ V=}\>flb^2wl~Uj@v:ٸL9inm IV{d}쐪 B# ޶ ,xHS\c~g$[~T ʌn3L7n-amQ3%1bwSz4λ5=k: ]'*%UCVY!1@+7hkk/k)ca<bo W:!{$%_AK&Hks)c6s-ox"0O z\ ;y1lr(W#y>5ZqR%ma ~ŇVn n<0Rr΃ߕk .Cl +ͨcܳ CH]͐cF:xCbtavJ“a'X.y̒a}ֲcցM.E媗"孻\u,B@o% Gf'aջK)7ǺtN_~2x&6v[O n,ၚ,<)9xdhQ(& qip @1ZFכtDS#s1qk0D8!Mְ F`y6 2.`>-̰3$i=sTdޒ(Ǖɟj64 ]\1wWa[6i2 5U7to} ݋Z4:q`cIefIsp eh0}9*- ^*s &n3*dmrj5lK[2` 4EnȜQ1A/J@-{{N]+Dp~#>(I !I.3 q,:.dh;9pHRlx l%<|~pҲMsi9{Q:OOnB^?(_AC9Zݾ}}IWt@>D JLJ^7M:w;amSwyUc\>_E[a@܎"˒dN7ccL%#|s#ghI&Ȱ5)t#)d0c8R:& {z{kuL3/(۬X!YJɤdp"8G^C|2A*b2U%kJJՌqk[w%BYiS>JDh_?Hm;3C՘\v06 T1RW}WOB%K*MGHo M"2j)I25]Cv w< z}t)#G5 Sv2с׽i9Ğ1s$}$6%N~S4Ziy31؋zHܳ;A{C+]A3fO=K; &)Lĩ}A向+AxrY2)33ZNɘEL% .Q5S`@[O\B$+,q!$eGKT̰1Y )I.GϏ NۺJiN )9&߿zw'zyŏө+gger ǰ5ޅؒ'<΅x\3ίCm.3Ŕ,"$rVE%JMƽVaحÃ-#b X=u^Ţ=x%mɔ=P:1v5^5Ȑ*yj<0+|ayQ^uyC91|0t^;v񩣞["L9:Q `8ly]&b Z񁯿"*6'Zi}2%j6F'%+Eσk 6,MȥdB(qd<{2A0ոm 'z#{ns!P816Ai=!YK1 !?ʝM8.7; Y7ּ.}*YN-ڜ\F22G6luhLh,lq\R'T2.&+/#]cmEV 6σf"bDzU@ֺ‰ĽVOgmτoSB %lgL ea}ހxD@K;fݨUR  晼3b 7i gJ=Ry;15jyC^CXjzt~ӯ6N 0f<<= bo6A7wP㼕ܡɟ&/>73Aɛ*cJ0tt#\5ՂwKE=g!G-... OxO?G#ԥ NJ fΆMd0MJ'/t3nVHa!I&!x##Sj7?Q1Qy8RF>d1+ڍuofm|`}u,frtt>S܉d4M>{-ߺIuQrLub=ك[3d/v\M}EeP*6^^n~\$;RIC+ wfy*}J< U豀<4*8+? A TY:7j+J=0,+c vv$@:-R,ʢ"bQ@g{j-W(!F &x NPi.בRg5HYzµFtad.5̺,B@BkQh [2Ԃk5Zݎgbof'QЃm Z)i9h ǟʸ{yT)j?>|n>T>Cp"HtS',/;OaՅ/~ywVD 9o6&>hGAxyٙɜ"j+Rk"ɛY&( Qu@ZPmkշW?򂖫Yc>~dpxmx6<o/6Z~ |͟&L=Y<MO.2_%J BwGNg^CNh4k:JaJL:`U/UYhaejjBeENT jUZ&HˈV2RК@ۥܔ#ID>D7 [HdZBPtg HuC?w'bAӫ]i\>{\F_|9!{GU$!?ŽӻҊ:|( "}CsSu;*-2Z*@H zh[~sS RoFuQ~կ yZchUF}(HK΂eȺCп__4zQWuKIvHe;[P, q"%4@; @ 3x(yFݧ÷#:ެkZh8qX/TZ2_`T;#p#HSSFqQ=s6/H|.M_0@ #o8EeG.v"dr&D!m\|l%ToQ= YDeoO"~D4iӦ7WX<F=eeFş 5Gt<1PT9ґ:/*_Q 鐙R.'BҭjH˔B5̩(%B LшJM;Y1i|VRaTTqO=:!d?o>bC*i/HA(uN-eV0>:ۦnb^-.~MWb*&Dw ڃ> cS׵3W^3Z W+!1geUmksf낡E*S-Zb΋9?[;xfZDX?aEI)|M[hqďsJu f|>GG;ԚH5m5E<Ց֙]&&Z"/Ǘ '=i$I4O RdbHTk:Nю`9 m^ *;W_;K3{ct6n0#@Vq*9 '.sT*TQ$< )ްƃ\Oן rq4yfړJwCG%r7COS&z:+/Φ+3u1HuwgE<FQp%؀a-uЭbw%ALp'C֬Q_AIX2(AJ.P1՞[I#F)I \;C>_m>e02J#z|J=#Y};Ƀk8g {XXmydzv!B-IiK`-VMtVJh*S#r41'w4Vf`GMF !BcGeܝjU;bu焞8H(E1=d]8!U-Tj֯y4~ AJFU&kfzvoK"މJaҨ\c+DPӨ9sd.E?H(٢Hk8CGn h3eAZT_/A~S4C8 F|[$L_=zqu#J+Sʲ+F[cbʨedbǮL< r݃B* ӫdOB5مzjU܍E 8tDC\ 6wXoc{Ok12ێr|B_|g>mQQiW#WRdU{Z]m9öHt;!IYڎYV:VcS7*f<,::Gӗv,GhŴ4yw<ֱX>[h^cWjD+9WTq ǴP'= j+Ot8mp8g {8KU0y 񖶼> 9?猽K>*̆ a ̙AB]ݱ˻"x.cTLo>V}%]tp&`rxV~3Ym77 7eMeC>9H>̽_f :z Q ̿cpP|67na>_}Q1VU=Vx :+'(fiNc`#CG#Q8!9ڵCA'!L^EEƄ0Wa-"@Μ@ pr9՟2.2A#Dj"p&T߽OTu`DhpY06ݘA ED5pjmy>(hD0]5Hvչ_. B #6-}g="؄ukqvi&NW:x-XX|ܡ(y$c{=\j`ʆzr,b #1 dRkX͠I4x6 Ֆw<1qߜ%1bhh58B{涍% K{Zbslԫإ $X),߷$EPA| JY"`{sGL )Etӆ*s9cF╝9 }kjq9%5%-79ZrU-XEMjz6Q9+%pTT$j9~]c̭NjR,Vj ^LvyB~G+E Gbh8M@D&**z`-i¼T`Jf;wr0;Y]uZG~4U@Qi/F&.Rpܥ:dx"dI]ub0͕SueH)~ֹŁ#HrTW!^C@\mc`cA[aIpD n,hc̈r-O/!nQ@\&녔FhJìp dG#{E  |J^! \zX4AдVcVUmV1raUwѹ+׹"YMP򶉍TA>)؄ $R&-8WlFd:%H~J)"˾*} X(O䯿WqD`iE@^!|<3⮗▣pI,p9+vXK@:4#)JHC1;Vn~s-0͵N6q%xBa[oB+Vn+yb ;O^0fY"/4[eޖD f屵0 %|ybQc:MgިTѪO)8iy9eHJ,@HT6<׿d[VbZMDb =Rad5Z@SOX=Ι'M)bHXsP`=jtP&Lj1*iu9`COmK- wS͐U(wJ|t9V ?*aϭ͕[9?:^n!YN;,o٠ʭ̈: ;WߵQ+[àgnj(Ҹ"LLjbu!SDC*-^Y©~$Dm-6V8 Wyw$";.&Ŋ%_4m= f/\vvkdFa -1Hh`l}nttJ]_\ z,[h6%ymHW5 N9VI>JHF$Kb{m?{+WQHU3;r[|+r,mʭn([7 N!۟"!ۿj4YJGJn~HAm&x3_loqwL?vF*ŢS͠nYk;>J{@ݚUmc u~ZRkx7X,Eu+[;Ã}_T=9Z0!b9?).z❾*aA8}_kxBzeO Kі#IUʐ#b)}.c Ec򒠕\h%A e8TgHg]Ux) z% %e6~Zs'& 0z~ -C&yRpH* D|M sdZDE3 fG цhц)[K5ac J x½0N  T*Qv׸$*9`WEX2 9('[E[G#){'RXa UF{9LI#2 hkcN\%`DE};2Y$t.;%%$B uj׉+PJ#fL-}D$q4f Y' (a2^瓘i$H'"$q,`yhB=M: $X%;7+6!UY; `Uu2,}, cQHBXO+F&\F:+Zxw11>ƭ`5~:3o~} cstsZ# v:f&yAsG Mk #Fp bw Kc)ĵ~>Їa(2tqt+`]wGb!qn_ ) wRU?0/;{?e{ODа=kϟAf:ߤhݛt8gvF/wϗ=Nkt_>Mip ;D5U3|  Cb4xBeפam̆.\&-n}}2ژ89wU{4ek?0Sgm Uq=5}f2O~F鷝Og酿yF7p3 l)u{&N˞`†'`LXs_kᨽoaŴs[~<`r)д} jot֯{(55W!m̷@0ǣ+ۻ~381H~=8%}b+ `,f2txZѶkxX{&^ٮ>/Edwnjگ}X~#՝~օ,OO/=w{v>DwEGMx&?;B٥qO^?A…MëODny3@ghMwµ5ٻ^ vXF6c_/z%/^ @}JwbM?g! KzMJI\,N>_AMeay~m:} S4 2:{4ϟED,%6Dcd$ofM[gov[$3keMԲ6ĩ<̮٠$U %/134Rl= 4qE ec X$VRDȒ.NreUe*c8Lpt^ Ef0PǙD]{[Ó >M&~+[1oMq "WߺZ>?C`?į @E%fm(AVDxN" Q)!C!j3Dx(#oIe^% UXٻ޸cW>A\]]mFH^%%b"V5̐"2X$r>]]WKW6ٮe):hӠmM^Oa,/NXSplN& HbjѨ6{Brf۝s֮̋}&3/dduP.ILDWgB66kJFdfTX]x%ko$;V5z\FFܡR^JFKRa"ՑO"@hCWsԅ&qn#EqN@YO!Sa=9GB#^c9h~1i?HarRRxE*xۥOKvw/~ -fşVjD9HN ނn.B[bw){l?reo@IeAtSA (qNL⠒ 8_)_}Ҭ?ˌ2 xޱnwݛ_|dw/j]r=*ow~˟/3˗_kEcY}He-mz--۪G V~[|/ϣz+?}ndGSдJ֩օo;pϫ{o56g2o98g>1n%{xxPr,zp0AfY~&N`u`R[ZU_bTL҇l.&b  ]Kw-ߵ|@hA@~ZaA/0 yҪ=Af'E%0KjP[~ib'1 v X)a~ g,8c XpƂĂc+Z]*JN=JKe}Nb=HQW@K^Htnm ]EO\-i U@ J٨eJ\aFYR%ئ>{ Z*BUV5;q)"ǷV1B&uAfM#AEc6upM+Rʡ"=}1Zy 18C&lAY4˫qrQ':7(嘣scQgs8UbN:tQ#j}4M9WH OQZ0)ζ5 )լ" X{pɫBt3SB_g$8# >ܔJ4w-R޵Hy"zNzisƭ8rϧD2iJcn_x(^Q#$jvT cJɵHSTfO7j}~S@ޛ(F1HP Ԧ,j=墴Ŋn=5Objт6m&{ Շ`oM B[pwjtBRSXkEb{5©J=zC]csu5`qI)*_5O+1q91= ]cqecu|'06<1߅,^U:TwzPwjIe{ݜx'+ۃSo !6q={{ŸL^܁?/eD-u<ܶ}JV{7$ו3u%g=o>D}Oӳw7L[Zj_Oζm;AkZ'҉:iK+ag_CAZrΠsMIH 4ku9+-9d3AN ;Qn:z)()7prCzM!DQZQB =xť,; О*XkZ ExQ !RkkO)K {sD133Yo "Ok17-"<!<7^.%z%;b*SA(MAt6 PH)1p] H$ދ#)`/2R! A^P;\J+! :6% -Kդ^D`UӾܨn,eu)^9ןG boug \DMٔśqQ/QF> lz-)9"WALs^l-,AŊ PP,=ym:} {M:E59Ν=t_j2[>dfT%ӆ*gꃗ&G"5Xpz=ay yW;ۡd]fu8(:?@/k"7`PSƉkLqų9P?iVbĩ% ЏA~h5q)Y񄻔Nlp+#{ BɊo ,y$19kHQ (rV2b)ZaccPVooYptLdQ)ZaW ڄNEk[f}vs(Xl+xub ,% \,~#\.E pɤ.k1mdM V/;yzVAUYhqEP|dIl;`y+Qҟ>}0zM9:0"ǎ[ɄxƱO9GzޭE2,wCYCk)u[|mdΠ}:[QYo%dܸЗjseX%ýMUp XYjhlO z;S'_U>jjxqG+9@OKׇ !{~J7/^(>؅m'ј %jP[|߶MdeR,Rwǫ'隃"(9XPa :`gS1pE~0Ig՜x0!gˠzťt!{o<+םכDQ$ܯ. SЁ7]=9[t{ В蝺"pdѢCBQhn50B5:jUXp}oڰ |rz].hIV4oԍ^DբF7].V;dXsk`O-<\cоVAzq`VkPW+E5vǏ[ʰ]"5YF"h >rA{^D5DB"XLS,נ Vly3{tdx2?<|kC)|S Ya}su`wdP]OygГ=^ s\-ZXˎDk|${ VɄď\5*G^R ̧%De+@>Kl|p֛b|޾8^5Igqrsl'N ԝr=a[^np}aO[kT ZTg,-eYe ۇgdTkLM*kŚ*l S>ɶQO$Mx,=CFh/\ hs+.A,:썥jGBҀ6B '1Tτ2Hz{ir\~>]e&GYm]Y!۵m]Lt@٪\{]G ]ui-`_Kդvk!3a ܊}C `2a՛p *w/uF(#/FlHƦ.T/J -,Oj&ybz|jrs%JCPX$߂}C 2G޻!'/ .j PtԦ t[gRxObciFQ{® <-+ѹuˁ}"&Ru:J bL[+3G',_j3053&5#h=eO5n '(t֦5TW\Ha蜀8tACgDcd])0  !YQR,L* k:J$"4 ŐX44l4=a[^Y;0*#'^:EUB@V3Y,ng *81A=[pSz,Ȟ-8pDM8Q:FEj!j6]n;m\(c"2$uT. hƈH$ i4x8tB *0m1Ğ)MqR!I5`4hć61dH- )W:%Fehw`Jޑ6[N_ g88;+'˺@䋸LgD3}KE0VzPeD+sk?-doqp)>Y=:9=~/ϯ{_~E-w)^07GG> W$zYwmqJ @YdH C,8{6H6{ [RQn߷uXÖ-{EYU&2X,>;DxbdS̚E+r=Ao՛)h&'fcaJ 1)Rqu]8`y[IZ+&(44<=tB gb^A=kc'1+&AÉ"5%׍%A .AsLҮ\X)zAyMNAyZNK͸ , qA`b/ p%qGF,+)lSkD9gc(˭B\)7;1KBr*yX:ȹomr 6z-NA5)&8Y(u!NQ\Ai{+燥BXPq]5u,_z_R E^F3 "(zRC"*bE`ǡЁrh1)RYuc"/\`\"$q`]I4Q@\VG{f 0mŰ}mVH3ωSC$3w)t ͇= ?8x]pJ98SN 3; I7ayrq$Y3@%8#9;)69k^WT+93v9TW PɊ<'膿_,՛H'쫳>,# jd)1Rq^7?&yGi ~cA.9Fg AJ7?j@u$h.vs:FOq[h*x2J8Dp]0J [SjМe=%GsSyJ-| , ̢285Ͻv39Q& y{*dw ,xzM[{fPƒSi%>`.OYd$-q&<^DƱv968'8sό5Ȁ\?z6G3# g䄔I]OңP;_]g]sYD&PW7[eWgYӧL+HId^6= ?/>2j0τ__?4 n{fm+R68V׀z&f 虶q>h7sWhuu~}ƗGi/(a Vo ?rWW7;S7!ժG yȶzwV%ټߨwUT'R՟HU]^WQ5Mp0bnIR7ɧFYZv.@d#*8 '%i su|V鴏JjꃎWV򢄗5ND83)\mƭ=!OdR~O&GyR\ W`kE3AyZ{L)ش/aH (2V" v~ȥrlphdQ,]C,i_۷hC.FfBmbUUlAq%GNҺv\L"$ҵgYڑIVq ] $ z\he/&ҳpI_Msysվls|mfzrcEUw+=UMW+U$nZvڙ"u=\ձn5-%) Vd3)cKHFs]Z[GЍz;mfg^KK , F8r\yMbAp%q^ ʸ&qvrA(J)r^AϘ ^ήϲOuiz< SHU?o#"SIG ޸6nnj.cfXw15Nsйsqk1&2]Z}Q_@'dB'Iv21QSahOWTuJ8й!5⨵K(-&&TGDPw~.IAa~f$OiL,p ?'w)x5Px`.?vƲZ!^,@ySovPrTq^{DPaPXځ] r~K&I() I8Psq!^Dy^AkS@A-}nNֿa#eCTX]rUȻKd"`Bct!uK鸖֓E*Fzrn$2''͓}.gbjp߳m!*\ SzZư)1d~ʵ$h8FHkp?>es5].s6& (p]:UW\*)5)pRWlcĺƺI؄PZNl!20hI:,jj/mOݘ?i5x@o88%EqI:R] *c$N}IADˠ}=A(γ1q M<6OBjm8IPڂb'WQА ثakc j%]IcRiݪk,'Q "z&&A70 -vFjMٜK#DYô%ɗo GI5-H9!xum Hp/fG+ p=lfB7p .%]s@d:j5Іs|搳baQ&lqkj5AA)pE'҇Ճ|&{ey'-TB@7Kͫp Fҝ:K:h˕iM ~kc a dAPqyܠBDiYb#nD$&<K݊ZQT%6؂`6]=)<`Tg=.zʎ{@WKY(ONB:8"i1j=0')5{Ж0e ->É"gBbީ77hVZւ_@qٛwBq zfwHe;.5{ %uUȂR8'( ZRŘnnPg(b2ĤjD3gt̀;0-ó7Mӻޚ V_[N9pct=G<ϾG*'ыz%Mۏ?k@ZK2P${b|l2MGqތ91#δ.i0tDMglu4M]z+)V4wM4#\ТQ1ڦյuɩйm=DԺk ڶIFS܌)Yˡ.M8[ Nj^_zݖl&OÝݪjڋ-F KǬG4l^ںky:8+$]ש㔜ҎoR;mze:,( 9?,Wb.ɾ4s Y˓kں&mifħ4>Bd̲u5rߜ^Iv|Ï=((a{G'~UŌU>Ɏfyx0u-"C :zq=/z]%&L>ts/g,.9rR,sF!ܞEr3M?afFmǟo=&h'^߿߾lDczU v޺?>pN^P&6T2#|]JNhFM3ޜ?ɭc{?>6n*OYSq 7V 5]G]׶kSu9lpGG4| (/)  _)țp>hPޫ-9i@~&OYՒ?Bs6&[&j{gn^L3X??{%)Zu޻ji3?~" DkJI)Iݣ.5u6j[ ;M;{vpE~Ž{xqHyRގzЊhĦU Zb(i "AFS:aUl"DnChI5ZK!(Qr {z: t^ٶm+N?śIsDty5@h? б$dwЙXc x_wu9}2֪_1C o~q|!;4 y 4:?GcK4Ze]M |l'hcŻɇ7bh04c;VV5 4TOw5Њ5MyS眪3WJI)p}IY@xSn֚+mc!G'>b(iĉ.jǜ nìg2~6QHAMf팤 Dmtw;cSIoܪa`x"fSL"@dB ˘R đZSKƏ#ý RpЮ VQN!X%M)DAȘt&r&cҐߡ Ȝ 7[-˴@a$ * ҭP\s6|? zIIH/fcm7 n^5vB揧?} <+n~3yY>ү^4:i+ABe[J t'^㱌h͛HAm5~6iP#RZn+[+NT#BRfhp< 5OJE=(LM?'y3*6-`Ϫ9t[ɠ,eo#!;\Fu*~)UӗmLKh N*A[kNAqKw: F;ڏe;٠́=s60FwsTfFwjs 8>Nx3~["[G1?H>T߽$T-7BuwDi JFj+p[ۈ"Qt!ޙ{;:FxB+@01U !a3O;K\ ԤtGUԞj9`Xwla3rW@-#; KeHʠCIenBz(UA,\|rBRIeDh뢊'N~t"CL gR&= |!sQgL=T$:Yä`zGpҝQs<lJ!s⌰?w~BNUV/`3҆Tcp)FLˆ2`hrkRq$ ́"XϹnX4R8qrN6Bԩ\3iw31D.LK1f idt.]^ u{%m.g5eBb_ WA;Yy!Н쾳_Љ67%ą svJF[8b$M@ YJ!!PD90"G|RƅΥ;# PDɃP^~0HԔʻjI+#ZlabeIJQTIf_gjFE8E9ĺr0&rvÞܼ2Jnp: FIyjR Rr, s_,\B)B$[qT9Q=K5OFP-J ߆ :ػ ׹+R0aK?b?K(&.RYߚ k;S(Wav&Cmȁ`9SuY t#0.o^%?TC֗Y8;1^J@@bFcHt% DvA/tui9ĹcA+%U-uOTR0 n͋ZARB,UQ+A!FͫMȭ$6R4FZ"Wc3Iyȯrk= 9]d*[Ǜ=CWrf(Fq(^rq!48gm&-&jY f OR3{/Yrr( l] j5)bpۿ./F4c :[?mT#,e=SOOnIAt7$nPRT-N!Ӽ]9":EA@.~mr Ke w~Z2Xd%7 7O8 2G_A85 ﳟ^'i~}Ⱦ՝jb< ? ^@}/L` <bSXGSȇ9~I@T8 G ?(DBb |F*{osE`D侧vWkW~}EbK5!^Yt:|ĠC#yS㎅"'9(EKu`B"#ei!A7V0?(bWQ6jSr;QR^6XMo/FT/Qce1BhzϞ|L~)K< ǽ2-ӻ^0Mg[ PRc?ᬷD_KqBMY/J-[qM^0Uܡ^D|Af\>Պ#1pCdmrwp@Wj:z5 pWf/q y I#)A@GT۳=61 f'l<$@`݃? "t~ OOXv)F^$jA,$e!' E2I|OR #bLJsk 7ᤢ`X4jVN &$/(}(unƱm]+FSlW;rH8e s(q }s8 Qj['lX{6Wj{Iq>r#̅2:m/h|fg"AC3rH[P ikWtV6YUXV΄ڏ86w^ڃl8*@qWKP#N~ qU?]VBbDԝi#M tr px45#ݙH3rUB5QvtΘBZ@*뮴@> bJmK1pbw`oL$SС;b+F_"jF < r) WEɱPlz#ZmHmYۋ(XTF.F 3y++ˊ"p\+9 * V [/PJ$}SQFkX>vSEREvx",aBGV-f(0LUTk\4_^!R!x$!v5b`jpx.?'t@sXB6⡠l>{bF !: ߾VVV>u%kUmnhfH^֛e5`M7baqЩM^eFL;󠰋v0 猸v]`b-[M`•s\ kuq%tN `P~كnRhwƗcK^oFaFs|*saE_wNv߯/$Gucc @r~ (9U bIqn=SG{8q:@͈U'A'n!7bw!fX'yw/__"sϕ}aHQ;sk% /*̩Z&3*paޗeB"@BJw{r[+[}Q9|9s-=PLvCLx$42'B3X:,kAC^R`@>b;;8Aa1H]' N+l\QG.ߤ Ac f[\" Be<ϤEJwD e/u8z|~w5t(^I-tuV~,I/sW>-FH-czp~8t#Ke:& ~3W+ PSE:&~D(/zzo4kHF(~yW/e_|ۯ_NQ߷j %a>_mhix~&R d%,x`z7Q,k%EMezX7֛(Y3׿M@|=ߩ4(Găw?g1U߽5/gtWC0].n`S=}K|jLKV17{r4'jH/4g?qMI}M4go|lZT/~s2O:=_2fw6ןÉ]/Zte {~&j5/W7dΩ2k5QYsޚ0ݿO^WΥJ&#W{8)>@dc4v! xD5PL?C<LߎjNn QVѕf{hv[k?ʺ0Mh* g:9(FMd:c.][Zc$˜**jgM?/ŎCPcM>.5Ĩ0} Tw?Z!ӟ1oLLu!T& D,rM%}I g JrXHnhQ*rL"[<|NC}TcR+D߱+L8aw&$ qNs^)Z]kT 5l[ Q!*zx$j΁5#][n r v|>` gjͯ`mW/F͒j{Kng|͐k&wyאVb5kt_V Wh- ^ī×MڊO%_Ҥy/)4:sT昶+px/N.ތ m=x0h~(ă^z ޾=ӔmRki/$ SRRIC/ԪW$͹ٵAp6l}b|NWl[x{ imJSc5`{LA4٭ !L`N6cjy2z6;:־Xyn,Nw oͿmg8i^wXӕXһ4].9fѧ(1UVȹP)ɺtlS髭_扛9CAP9g  `\2;%`ܴAv>Y^+jx'%@',YsTX!" x 2ʠb2[Q=SXk˦e:KZTH&OÙʡU"ld1]˺jZ+ kyd V/&i\a]&.R@x㓴Y HeTh& ш{7h w78fE~(\T45zYi `&VĩVFHݒfzrd*D 3ǔuVY<ݡC˦PUΏ͍+Ox4=N j%ΒC<,E8g:*|UAI 0 5yhhd"[J##vw=$}|ӤzA`n< !Xi0c͊6>šWB[͝7ADUI)]M֚hgBQoHJhEM D]ȉ;2 akAd[#1ѐsWm zzz aQ瘹gOo_\U)oXK/?m, t14V6E1vtK GyVyY;9+l#2+yun6>҄0PVhfK[w-wy{ݢS.SSKpĶ${=sͪa2#,QGSȁnChͶseO/?Q)FPF%*'9:ɒt֪ϸn ͅ6[M̶-ͷpOw;H817I7RY ݙ&hS$4ӿJ_7fe$1|q>8ןKq4x:<tP8G*V!xDƷ \԰+T[=λ@Az{0d\#Yǵ+kh쥨dB&5t}ĥKS SAԃ!Kf9%$=2HGE66wAab%3JFLЂA1O wYɕe/ZT#2KLC6b4O+7Mq\,62T-H-ј.g KFY29 s@_HoGaa| ;«=:\fy94銎ayw9CqVYoh<)F,<V>c z,V8K` D`qNcދWP%&/01_&c?kyAhf{=U˳5ayO'0]M݋i]cB|rtZӜ"GEf)(9||0ȣt6 1r1>:,T&LiۋiG ݟ%B. N3H鯿Q !@10[N$4GfJr);|6sB HT)dӠТhL1X>'4 55de9KWRof_ږ C[Znƨ ϸjS–P+bB8 Z|#AW +Wب>PilI!['hJϓB# M ш4J#c evlu:)1 B}eJOljj>nzR`5l=U^YRTA/].@P{ёy#d;_ blәqSK\P4?}Rÿxq{M-F ?զMQioˍG.BR(?\`/H3_ߗIm('S.+6Svg.K|y a2: fF\:i"@e![J>ԥ7//tmDյNrs [R̭t %YQD9 +!F As='G2 &aK̊Lr$]h%aq:x,zbCV}J;!A0031p ީJtG!PLZG]Xi7p-(v"O~2#y+P$S WćJƔdRg+FqIGF|>J`CoҽX zb1ݼ 4D/H~14 cɍE!mabz<8ܗE/uD4ajdmL*FS֦䝬Ms MkC0Z D{kC85M P$ 51nou kgtwkwI!&Tjֶ~-aIFo)▓8-$]pSe`1ܢ#٬*DunG g9ҴF]5$aS%H[4>2QJ'iLMPaUnJQ+s1$ )eh !B) B7 {w }aJ U{ܲC4Pb,YMUU 8Yౘ tI k}VS {58a˲:У 2:jPq|utZu,e8lz}RYfD+.D*S'T++Fg~l)A("6L߫>fwSZp*~ϴBfypRZ?)Ə)2WNC f&R[tUEy8qV㻻UX6^Y Dƃ_J^u|Gñ_f۳zB}RWaڍwry74m E>D7C6$AtɆ ^q}|CTCcb@%dph Lv El n=%Rj[t%fb&H 1ČV'`sʺ C r4©,3?XIΘoG. ?VWъЖ[rSbS~QmVYfSeia;\)neMbT.mƍX>=V`ĸM7&W4M B H"KS?Ố;8|a>Hr>d3=j4sQF㇢* %V LF A&V5RjdG$)!&Ku-̼WpTSJn)n˜Gͣd6~?}a=\bZ<y&YQl.S%dSyNߊ؃+~}=Y\86ATX6=gnyɑƌ47׫)`LN8jm)[:y:z֢Z;zvԭ긄eIZ!@ƄBs.IwxBxi02%*fRX%&I3kݛ-B1\3vdž5BH}, / Ёg\?QU;GqfgղB +*w1 ^Wd, ֥[n7h)̜]J c!/!hHcfD 6&>BVʩv+]uٕ \S,?d6jت)@rGqg%oFVZG!/ٯ"Hxի&UQS:; GIG%eۚ\-+ Z[_늼07ޅzf9W5M/1Ռ7-t.7?Қ+tH-!^-!'gYo)\!-Td+ %wAZn&$7C=BH}/ Q݌k5&N±فUqN;9[ ?ʑ33~ꚠWy4GI" 騆R8q x)ө̲W_HRux~>p;85N1:)"L1@~xoEqV։X0b s!RKm]}rXWcEUf$dX+Կ*Ёg\?ٗ 2*f/m*+4o3ړZ(ӇS|}ܼܣsҳ7z>hQu 抔0M /Y#xĿ֛ qdS$RQ'A8m @Ao42ggEfS?1-|q䨤mo?ƭSe.K ?*j|E̓y5Ŋt8;2lT )2әS.T;m32S.c>L|ćj}})b4u󗩟*<f4t~Xa.UQNa}H`JHV9rZqoajBJj/!XnrK(UāLH"Dh8V:sdL5DQڎ6 ;=&ցN"rMjT)v,N4SbmV=.k-Ъcec8KQ3k4}S%)٫ɫ^T _}Չ&<#@'{"=BQ8Kn*׊+ZCxBu6ISbmg gZS! Rʤi'thk.̚IUT)lҮ }YsS~B脕%}}~uT1FG'-A5#T,,\SLIqdt<'G\"ՂgjsrC~^@P)NTabK~?uX?0j NzYTޞ}_/!W5/ұ 3L;?L|Wo d?tO,XU-aZ0#D5qAqy8\SʻF-58mA8;ǘ~ц=c2yvKwΦՃ/' [Iv v.k;yBp*qGoNuFGvZ+Y[[opN2#&G?y傁V_n xE8$BbY^q("m7(ԅ\*OXkz5(U{Who.Иޗ)L5ИrW[xͻ?!IU 5Q۾53¢ю_E?^">QOg7>dvqdxBZ1PLS Qe4MylmURfp*8eiN”SɅ ^ɩd@P"?<^P*XrS|%_ydULN; ֙?ѮΦE҅m}ڕ6 ?G?lfy/5g[UѿmAeeh}~uqtogɑ.[Mg9F.}@cMϏ^-/p*\~4ww_h-}.[>(RL~}#jz5F58N$[@JF[mUdЉ$X5OfI=*2<%$),oä48Bѥeg˛Q0CŞT -IY[BTpÂh&x@ִv y/q:`ɔ`WT*dB8%ea=õEU7=}vI9g(K]*OߊSJ_Aq;lŽ'gf* ,18Ubˬ6(M(NBWH<FY#|ʬ(9A<9:檘ؗ_wXsW(z!3]+CfcCg5 ;/ߒqRYpH{r -]CSM$'sUߤ=&KU}[v%/\4=HdR[aS7M 2B7**O몿 zhZW5; (S)Nxśy!F&=Mlw,<KsAG?F˯Ơ(gωqڧa6w\(>!ӳ*ɒee||w +ݢ#;LF毗cnU0^߿[P\G ;ƚ6s$&c1 XdqL2%J6J$G7{'ykj"zњG+fPW3KKÔ]xUZ3d }uS( @X)⼊bt# g6c)g$qKFnfF1h"I2 A~*`و]b-F-p4׆ZydI]R6Yh]5dH"0edч4KȧVJc!8c7B OY}݄]?~܌oq~>(o)6}MsߛE{|wv`$S%!$UwGxgzܸ_e^ػ88peomk=3H$b[L#[ZbUXU,Gngߌ=ڬFZCձ'f z|@E+uI6|Gji|ZVjVUuS׉{aN|wU5ᄎ{x&q5Ǹb=ؽp_){G*J Jޯ;3\:pj-0959-s@tX\1Z]zN"*kB.g3Qق{49hoa/3gƩ Cuխ,;RT hro|EyubNRs$ S%f0ZՍu'V ?Yo1Nt)@O6nH|w6X%]!FdO/Vy7ɬ#WvlD'OÁᓴI? Io%<`Ԅ7Qh.ŸG vwvwF?Z|9 ӆM3nڒB_eL<F2/ƙ2G^m"RRP`,6#4`!4#R^DA2-yp#Rb{:bUCҙt>Zmi3 9&~k_0$z@5JwhU] ym, V09]d/t {pcɰ (6 qd"Opg7Jp3ytmJAShuyL,GZK<A: 7p9s`9ohn1y`&H$ ``0"ō(4n"Qq{x/6Zb )pOR8n-h9+vK <2-1Ro [|+{MNxMblHeyq c%J +EɥhNXEeWL(us8s+ B]ɖCO#S׋~sl9%rjiJd? ߊ*К<~}wȰ_&=Ʃ;*}GqOLg esՆ۫YTy]W΍=a:Χ~|w'ц\K')Bsjpٿ~#j-W;;0FGjНUMWR77@6芩 ը|]jl~GyJ:X\4 z߁--J ibL(!G`t R=xJu1͡&OR9O C~H@fTM%ITmkQ0[Zx&qWwc&nq>,ݛ̗{b~m$ǐAOkR+*G?MԥNi< H:piaR FN'6B3:=62ޯʩ"RۂSO ) A?:_7Lf-3. [)@83 $*o_UlE0<"B& Oh?\՚i7 d0>uoo5wOcM >Қ_oӏT򚌪*Nn&>>po(Q?) :Y)ȺNAu=ZxHMJI9.TD9¸k/x=J]^^xirpj7z4 i܎hhjѓMc>uI.olyY 0{f]7L(@ Ž}>}ضC!^YZGsv,pnnj޽8ɾ5!tk|d~(?lvoC/ޟtСە{*B@+sjX6ݍD.K0Jsf0wc{64(ϡ;N qܻ%) :fgZAl_R$K&Y6&I$JByq[fɥa3"Mq],Z.;IB9ZIBՅ HYp%Ζm.(gd4&E|}/H@`%2bT3.!Uq> l~U$C 7oIFR hiFf?/Fm|oTOP5 ^%<9"16$rt> -ŶZIn znߑ ˍdh͸OR vԞ(/N<޺eY|az?XA=gFMsHgɦyiXލ ibvӯw~r8}􋾾.:UC-f=xww$l`<6\h?Ng| @"cn.d<'G*Q* ]{)0NRmC̓fe=m:uġju5؄4Njn\Sgu'bkY+*ii6*r%.- ɝR @j>hR5!g#>U%4eTmWYrԜؐ1JY؋Z(rӮˮK L3M{o)ZW2k'1Gk&|ѷ5?r*9Tf(eQʃ+ eaH?̫-ʜ!.!'c~>lFQ (>BJω&ZEK16殀&qÅI,V2GpGcgy׺~/#cJ:$V.OZ?9&wK5{ʄ֮w]p&{xSA{&nΙ? J^1]թn\*$Z!!u0PGm'M Nz^?0eHbFbM`_hp\׆)atXrrYGY)'-%-ɝtQ^B I^ܪd֤ ^.H8(>uΐ ~r~;Z+8<'vynNserB"KRIQzl沢#):sYI}HJwϯL L@kD ں"2SW}_hTu"iRMXgW^}; y\)zl[&?} `}|h+(A™=ZWOG]H5ddjO iiQ!ѥ< q54T|=R4{VTyngTp?Q.kU1}Fe{$FGˢ. iƎOqMkhYq./T,izZ¡e,zjkX^z-T8Of!(JeͭA/OrN @T)sڷihy>I,-*hl!zZ^שuh- Pd r *H2@kkTdŹ4F:P'7 x#FNUn@04v`:f&$se[P%Q#(_:k{#Q < Naqmr[22Oc#1z Rb{p4.R!xW:tIseN(*/걖q/Z\%d%R u ,r-j)B AGyT;rNhyY[N',ՎGqJz(+%eVH, .`\eI YP%Q&#YIN H7m(VR _AR5FtHTu0 Ͻ*A}x"@< t +.1hDR.rp6,i‰[!/tbﶗrNCj@ոEߎ:$'6.t_ϓTUBD"\Oc(LZwzW_gux !'qoҀ6)fPl.j Bkϴsri[@9ߪ =[ky-e1aSf9j0UgE^tSz=^O/Rs{+ϔ(‍F.??Ԗ]F(pbr?(b# Yݭoཥe?VWTwN nucǨ<%N+5U O?6}:.,*JqwVy]+%qJ#DS()8AHÕ)5<%ٕ9SRz5M޵q#E/^O&8/8ျKonәx;W4Vݒ#bX"TA`eܟֻ hDnJMe*6*( NjfeyY))HY7s[a-w8Y3kC>Hb-7:d**7Z^.t:2FRphHwL3<*("NjI:lLùGVQc#.mT+wE[3FRqZ9phfSk)V@Ë)\dü!J.%wKU_'餔mv%<5]I%q8J &jhy<ϴ/8썈޿YP֮eY^grݍƫ_%,k8ɼoW϶i땝{Ծc,?<#h^iZKꎹ4[3w+ިXIO~^SJHB"Z#S,?VAԾv;:;P4>; G0)ɔrt4,X#\N9M 3qT.4ܽ0qЊk>draP+5w he&w9K[#%[@]^ 4%SzDP.EhxYC.d*$`9GKтZ[G$~BJلdӃvqΐ-]I*MZ F¯,k.xr'Y9: {;,WCgB`M^+tgU[Xfz:ڳufaYZE1 )!DX˥ԙeYk]|,Lx/;_=|OPݝ8*WOn'$)kAwstu䛏d2G<9zrCv?(JM=]]=QpRdIF Bp4'}7*$Te8Y٨C0'=O8z? 'EcN(/tFd|Bb/ u8d]D݅Lw7 a\)]Dg}^*Ş+tiv*j:N#Z;fյmՏ5_-mg!7\zo 3>!Ȟ_7ؽAԖԿ~. X/^|ëB3߼˲q&e>$:V~::uvëbTMݏQ'.L+du3Vb?Ft@;r~=4VZmΎqe w*9[;Wͬ+'˝m`W΢c>F*uH*8:kb&Z0dPCcf1F ֱ-6{arPyX=᧭C걝4չ,L6Vojgrl 83q øz*Wmk'/bً6G޽:~ŪȽQx͆/7o|ͣ$-ǂ?S.U@;Cfmn*co[3M5o.ukH ѡIYiV5R۰2Aa9r*lK/xݲtaEKxueI%̎cSM=b{?8O2}yYUsG閥| {T-w3G?ϔDyyTrES{7އ"a`O"a. Җ>7ϯmVCach6Psn@{99l~ԭHI'jmDRm˒d ))AhCe>JwŠ:!#ɀq婣'sédtL0UlRವpyuޖ__/u˭@}W*Sa ɫ)iLKZ̗x_m7}š?jn53{{{s~K!ϖu+{x}-3Qݳ\g?~[G=TQJT1B. IZ`k"&{|c8ӍOдh],mH}nPĽ#+0]ѫE k!tdl0jgb37ks?3]hTKs09zM#yM<sLjæ *{Ky6 cGl 9% 61B)uV M`֢"8AiJ&Q8@&P)S2 %tܡur&TYIrŋY+ +}]1'/[6i`2!8.Z]kdKyL?~u?I`4,MlR`eF>}䚎-,$t~c=}Og( av3}SHſ}._tf Hso*BAfgq#: >YٰV|rB4G@h385̌1svCZ %'| ſnSM+eBϾY 9-3`Ke7l/K3?ڸ8_Ą8M?绯|}~oa4jBRDQUkgLYgID clGT_x^-~F~{6{p<OSg/z6{c {aoŏҟo?nB+$!ND+Hɂ\sH+o~d5z[mW=KΤ5'ܐ2$&1%H$3Q㿉+$|爦j9iэ*VdRX (1M1HU 5!Jd8.O>؛Ÿ ¾,l{} 䊐3l _`r@b^ƽ,dvaaμg5p_Rۭ~1{?lA5{Jo鿲kB-Gzid XEzθ*FõOyiƓp3xbFXEXx9CnFS`w_QR`ޓsКlH!S6xI:uGemH"`w?mn,og[+nmE&\ ɹyqrK%&7͊=%'~2N)9ǭ&g?Źu]a٦9sbGAmYWO'E=9Ȱ=Nι=4LBMX+2QRHsN#" FaD )m.9B20/Uf2RZ55Hy`yf9NJ鹴.Ze3SGBtJj,L*<ީ]YX7+$] 0BA I!s1?5Ɓ !58JcI  '7QN vX$A4s7h%jvTaNǦs5rU%$p$#Fb "U`QKEQAX gkk9C\ y֗e" .p'%eh8Px!`!Wڑfd&MJ᚝Ys k <9צ@PAXd3Pʨ&CJB{ ZS]Y4D:/1`RQ!"`\`"{y H$O9D䪘Xw,pϫc:D1i雜l< $MTJ@p2i=I /2ܽy+K%-UZ\cw2IS9OCI CPEh3l܁K6j7݅h_I r%SCiiL v7QRSvNzhӓk\wn6|3$o`qʩ}An#°wC4^E\22$)`!HdwhµTD ",DB g4:J]gz'۪[0'W <-ʈD&&cJ"/@L"`C:TL$!G _pEID3,%" mWh[0E kO㗟+JK#ڠH&1^C]ޒ%(`Fd orS]Kgz#_ibP"ie78A=i1%jc;ݶEcdKFMK,~B(IrckMOO^{zDA%hLACD[[_8.&LUҢmNދ߭C ظP<<}nHA?{jJ?r2F]LAt:_2o=oCDt}VBӏ͗4ӌB[2K" -kPKZ4zt~&?`+=JN*kp/2x9ӢݹrŮV v:1j3Cs>fOw Cx}Qd9u g 9YKZmCR֗!k\)alF_ӢwZh囿'z,GOtr5 -=f5~I~ah[Y}܄_ kGen$y@ =0O+))̹ÏJ4]6\ENTZ&d %cqp^aF0κt9":{Ly+uЊ!8똑=oI[ɭ%,|zӎv<3g/Ùg`klEGN;)hmqDb##227D ȓ XyV-ft*Qi>d jm*N=Ҟa2}$I !v&%O1vIRX$Ri h`Ce!+-ǷnhXL-gu\xnE㟗|UuZ}a0=\/I<+.huq{.VE_@/VeO0d(v7<-'㹳E ,SL32%teKX(#$vі!EԶ 2rb0ADE_8zkDu$clY})ߩУ؈bGwDw7icPx^q!1\^~߽C1I_Pc>zqλ\Pݶ<(ەTy~#9{s+ A='F -Ngkec6yVe/SR+:V5ZAYte*9Zjͺ '$2gWJ #F_ `SN.!,J[TΗNq^)=0`%JfQz[s թVĜUs]yN<[:S[G%:WLꤠ%Go3}`\ lTI @[OyK]aI.^I7 DndWP\v]2e++]6?~Exu 5ңUNgKY:#w6(0^iEiS!Rr#La:G4:*ˁmp5b?aY-!Gyn˂(|#+f e,+:򩪜L"o!!Lۛܨb4p#5^g&43663\Yģb4>X'XhbT%:1C=E-(02418&fUZn%G%F][_?Ҭ{jMCWm!5{ ~Rg_h]㇟^%"r.۾ueO~tCmӌLJ'AyN gS͔ME_䭿liNU;?tsZMiO8Tz yMqRBUv 4꿽 lNkx0X\/[^W9-*oWo.$9FV@?Hu4~ Ȗ|"Nc*7hV8.lr57P %' Ҟ5ai1 ) /bm 0JiY^iTo^erv<2Vk#x܉a=(1b1m􃏏! gLN-IŰeJ(>%8 FrֵbgJqPI ^%e:)&DѾ»R |+,/鱀V?7%$k>-W~S>>xaХ2ldH^IX5ʰCZRw ;`E,H|]moo8tED3l+yww]̔qۣpt _9+eD95dBf;A _"a׈)c4m} 5ѻKI@=Redꘟ%Qbx`3SC5: XׂY}$iEQI%,/le-A/pL ɨ[Kh+Yw{{X Eѷ6Yݦl˞DOL IͥD"sYZT39EAdy%*rYГ\ow%ޠ7W}ba3040z.JVBnn<#pQHiDŽGoAXa0hܔ #b_]07be`:[t$&6UՍ=UGA@C)nMѝnLIUǭhm8Miw,|"Yd;xٰyGc"5hm:6zQf6wԲn!羠T>f ^\|+ikG /["~XK&@MKvjl{vL!iޥL͈}4K0IJO.3t,3|w2щP_\nyw"sԴ #:P:f;+uU#V n5!)&;O?) 0>y2TPieǾ G%+9e}۫>J@sп"Sڲr.Mm.L4.tPBj]6k# 3ՙu;:YT386Ke`?6w`eڈ4PD|k<֛.gݦfqܳ'OOϗ(hGkB1//.Z.j9;iFB@( V^1jFlwydX -Uq`;e+ccM=x҉ [53M/l>0x[8U2v=2wւjjP_(@M3p|*A}RdXP[7QRSx RtR$ Rs&7O d3&iJ7/ p=ƛu9חFȴ9ңow 쾩ޗ9W_KGy.GzM0͟w9?͙wk5gk06 jI"hH۳^E(` Pt*: NӅs8zG),[i.`#,a+`Zf{,=[l[[ -G A% Ƚ駎d^G*H|?(A6m[NPv|p~|3|njpu75ЛK3( T=;k#ݒg82VlUCđ[ eiz)x@#Hx6 < 3j`bl-0o͈4SI^ :3+-d+¤G Cݒcy2ަ!=&NaG4h4(s':fj0T ;D ƪiP~?3SS!j 4\р77w%ueE-ŗ֭CM 8Z 8͆ߤB687w78Ɣ#6 ejCX@Fa䚁(.4o]dP֟?/Ɠ`ܪniŧ۝]Ou|B pRN]^ p}tCѰ')EY/"x/ 2P1plG/';W 3]+DR=a_q eH ǝ#}±TJc3wCmߞ[هFcN7ts#X2vG;_mp7|"C3DYHΫ0LwDghh=CL(%sp_"n)W4a10C.8(cC^Rϯ;87-nB@2鉳#}%<]D>~wc9$pLR #_={({PBXjW#IߪcDzx-6K[٦9eme#Q|fvZb4DE% ng նk]({(8|n~ Y<re RZroxY"Db5-NҔ&^gbj$o*CnO"nWM`}큇_1W{;)}Z>W\> Y68-n#/cGjhш|u1}'EvsGfBvbf %kLc.7&i[^KW]t;ar?{^|]zpRD>?\<糋w*no\nUgZܸ"̗ޞd!m#_n!4m+7#M$$"i=znvZΌGŢ nּj/=KAOl6*. c|C<IkIؿmy Ƕ6#}t7r,% 5oG$9oZP6w\$9Ri6t&?phKB}|ɋ6ӖdkZ1X)jJ~%yaGb^pRHg/f?)yeңO P`]wT0v$Bz*OIM3UD1^2h>a,.$#"-u\O{n'Q2\%lBNJƨDnPqQ꣝tAĴ{{| nXG{2ZD)LkVnվ}qzcXћ}25uՌ ;z-5:]ŝ_91R!cT00ωTg0vy>f*׾8 iK/Ǡ-T1(Yה݅>E.`;^[iLo Ү$RKs"8!CpotS17J: tI~Kʣ ;yIy*@'8$>5`\zB̃K-d@M /NLa0$P:r)Iqfuv]S;u6?kGM])pD-W-w%ɒ!$f_PA@=[ZAI fap Bk 0D-헍`Fմ\ -n'?84~9,d?oOéq8u<SǛ'f WCC#r%S1Sk=~`M?C>hx~=x?V~۹g׃pbP3pٯ/;/_&/K~y {V.Pn*a}(`ZAfp|"BJu6=!qqC{\ q讌sMr Poڈ,GRӚfb&&99.'(2xݕ4ݕCX RR$a13dO<,\u5{Lq.N<,g4c͌kafȹ|ƽ#)*x/Jx/?%/OoDI))J;nP8lPA]8MaAeLjse@sR>$Q3rH$sRt(iVZJ@)dj ֹ’d+IFIPqVhs ~ Ϩ JKuvb} (N${ ޣx& $4##(؛SʝfkSDPL({Xd |q`24U72#JTci?7tb:5K7T`HYAP##T+)Pb\1){R:lkU3f?:k>hkcG2c9ʄϹ@Z=W4kfVbNMe؉BL/iw!̸T?084 ΁A0Y xRmU*,P8I^$Y9#;gf *Y,n¾a 2SZ(dDԌӱQ~";mS@@JX G6\r,1Xg4GL)1kFZbo-1pRvZ;;9SMI`j_,s2u' IAD@IZŞqJ,*R鼏Fqa4BDSl Uf82Oo< Poߙ?n'شu.eEOxÂTQN>O޿~~ShX=m;osC`<|$b7 >4|v2 Bxwտf7<; V?L>xD@N0#ܽ#J౬v~G0 ߞq#(܈MdjΛY3b hr=hr"yٲ&' R\G?lnx{ =,|ޮi:qsF9_GNE8"hd_cv-S(H)lBxud%FPiQp❬u UꍠRtۛgVcpǧ+ɮUlCKIyWJ_-_=~6SOtc8#3q6LgnUh*?d/Idp`е`sڙqKd~c ?Ceի'Ae=$&}t_uI@sܝ^B.tvbѸ:&}&__ 6IDO6I36IT4º3Pﲇ\ GuetpRmZT' @ha+Zײ*27=)T0*pKaL~S(5ZEe Nn(jo pr *SwxbCaTSy-Q}F˜ƽP2ϸ[P;8nL;yNHZ3BR;*$l~NH-B%e؊$Z8.FD#%Âь[wn, ,c~4֝{89Z$^Zw,j:Z{o7~Zw=.FJdu{՚(.8fI nat*86?rLhJӍA!K,"t1Jtc;Gz ]pkOuR`2 _bYj:GEvk" ښ.+VG!(Bonu&s90gMWWޥtES!/c^,tj1&[ɽKŅ .GTTMWR#-Koc'.sL'Q,Ԣhc$.s>Jg}a# CZM 82YF%PPp30bCe^AHrXC#q#)FͪheHc0aYSRMBU4u*wme\F)Aa1^X<܁KjPb`N[d2Ma9 77G-v>^z%9Jr AڄZ2!A!{k% P4Yêko^qX,#ZHqՠZY0ǭDBk7I2kCxJm٠L>ʻ ]ys<Ӭ\ _<. ` 7nK{?hDa^KW3``7>,A5H*ez`a^E VH!ç>VsTӢ`%CDž[[T4{ 2i>ϠR:6G#vX˯ *1XǐK.:bt&6?ʴ4$]T&iF5~$`rsֶ/niFOQ6U29A؉j'~d J*?w'J`x飇XL>O@9CG*oٻVv]ٕ|%Ux#CP5IE\ f"*v]_tUܶ\n$HcH.E])'{ɐ ވ~3ҸQɪ<]9H9׊3N g>RSQ#FG\懦L]9B㍚Q7tAO/]т=w:1VdrvgVc ^eԲܤ.B{/ZK`*UWZE|U- pvnǎ#U PSP7|/$U5rͯ?}>r-[Tb+DE}~f>mظ[ z ?jKngw gIo^futR6{ByrB=q"4F;I` ]LH\2ԸI R6UqjI);RbReToo?-#CRSg?O6|v _foxI8#@Ci>d9x>h). *-^tĘwYXn;j3Ek*y%XwDN2ɗCPs5t Oҩ#T0y<1ŲmJFrUOrGq_r<M`qZRw۶J4 8RV(o`aF2ĭ˸Ou 718MIRxqJ`K3p-iRU==bXIv 6h?rQ%MucDV ʑ'0A#g:E˟&U-%\D[5|"qi籗:(+K(* VE #~/?(p AU_s" PaC-&8 &B)⢐ WoJ [~'_gSO)r)RQA[N, KZ+jX}ݗmm$;͘ڶss⥛|GDֺESDrbN?{۸=.'}a$IN^vaͦ${$ɜSdD6E Ό%K_]n]]UP"®z:ثι% WnX]B<,#)uaMb'gZM@ 5>'HrW***^95}mpD'MT@X9i0iUBG{xfо ev2z9gQ-t!mObͮ/V5VvEe#/UɌ2PSY_%= t n6ُ %ly▰oٶ@RiB޺ nUcd0= ?̿Fp'GU5Օ1Y>jh/=n:;@ޓ -[0K"ZqLx2k89R놣1:HPe"j8aAq+_R5ܬ\ Ae5UJdn6"%oPCMv& 6΢Ol` 5UF7D`._0^VS*hz1rNJulMGf8GVbIH3 CM44;?2»!-O'J/m٣/ɢeZ> \M5=ה O'Un5 GC 56LWf? (ό؈E"8dd8csQ̗yNɲ瀫 !Qg!}AbUDDDB㏄4Biœ`.| ;k<{7 *=?L fziƪ6)<2>-h~+Vd=ŐoGU;ژ_N8JexS8o$d/uGDcgP{I6q#/gl Yu* 4y̮gDkn*pΚKfGGKnk<Ul?/V3?Ne4&LH!8k~-*܎ޖw. lXDJ`x [0iKQ( =jM$A*/Z/2Dk&DP+iF\p^X9trLRI)`2 +i \YyY+ly>k*l@PنSe% MOёp>j33kӽԹiaִ0Dn+ȚlFNUc+m/7csU('_76 4ԡ \6-9ؒ;']W\=@ Cޏd\]) i{S8(kJiԊ J+W>0㋙2#*V?nbRno87p>]+AfF7ÛkFY"jWz*hDj/(Al2 K/g&oYf",4sA0A#åwJ*I+o'gM?KISo&ƯZ-# qDb'\m̃R[ΫxϞ ɭ 6Gmr4) c=1ծ⎝j; }*"/3(Yv;OֻO9{k¿f :Y)-< 4L?pot^ DyWIHb$M̃*SbH较vF(M9rʉ,!ߤ~ FrҴY9H:4O>ZaikL$VQ"M1̞֭"v(4JZZN8xi6sont2 4GSr9 oRGi A5nn֨ O4b|xuKXgpC˻JkL̑5s%1 ߺաn2 ڿ-;xsمVnxm dClOFۤ}qKk5QEg\n0RC,%1E %C_Yq{UDHݘѲJ΢h }1fUS*{;'E'bV^rTb(9ux8KEEğO;pY ]//W)]W{ hZL*mp"uQV2pT_F4FLwPA5Gd0J.)3uMr%;wx%3j{&uY7o<]EY@|%/~힆hԢ ]~q?x2;y"_W1FELu)8V u+ K]usz єc VU!VӜl/s[ ^ 9oiK8aM>GUF4VRIJ]́ p:f!%P+,q2KRp%i`[}>-pO~q:Ec 'pr;@om0]`ԫ8q[GN @A&̦́vs8>uH`%HMn] lt[WAR+tp_Cb=}#j3FI%d1 t7}."'ڿ|USicqM42|2(űH-Θ1=Rp^r{3;qD'MTBSX# (aҪTIr $7G_iQ&pL&+vNF=4-3|t!8,ү7zc?,<Ƀ ݃qy?A (>?,)ϯd_*axwSxx-4",zdW ~+|ńG#U- yS( e}?s韏 ?'pm(t,wb0x@aHupVS,uTVHS"l t zaTk)[_@b1فd\>+u|>UA)X`=/T7Ai$Y| =k4ry{Mž{"<^  < M035LmH/#?=8`gnߛy7a2٤c?o@oӾ%8y{yz\{^^k|P0>ۭx[\XShPn㽗Q ".K#?E\X'QD41oL[\;V$C)= yc尧n9$d0Oڟ'NU FŋN'J&~y5qMP`, Lɔpj!q"*FSUNw}6))_/?7Fya7FyaTl"j(vU-SQBbEp>}8o4j3΀)Sdd$PmLrTL4 ePùSPMWm]qhr$ce*+ KL2nq*"Όb]]'A@i-,ˏbǚKcXJx8-BnK$Kh#S:? !ipGE=Ѽ& Y`Tņ3*3 ʀ0փ1O KqpE~}ݒMD6>n7aASt'?QM߬u.D m|3^=ț/r_砣Ͽ~rA< o'y;"\¿ N!SmU|ܤ$" :V&s2!H̙MM0ޕq"C}q۠I$i .ɵȖmޠX\yWW=xp'kܞzNmAXՠ1׮-$r]jwZ HU{fm%|XSs;' s;')g)SYUBah'sg `f}S&3o 6aprs_G.9W G }_ F,YOOH喢֭ &S;-'g2ʀJ@Hqi"0%L5Xך% Jf1),ԸDxO;:7f#>4ԙBN~| M׶ebZw"0iPE&U (rZK)x,H(DcfLy#)]I簣e#D+[fvwPkt±xcJkp]eBx:*%f)+=R>TK?P(M{ώV"C# Ν(^]ɭjoaә E5r`o0T`M%|YOhcW8R iu,X- 8R0gPyMuc0!`Z*C$+(/1jh1.фYOukON+mpJ05|DWrgTAԉ`$Bx|/<&ZS9''`q$&[A UL1T„n21@^7)/6G"7Iہ /†."pOޥ{Ӻ98EcINƎ~RPRjZ.MdNzux;$b9"_Xa֘" 񽾿-YH-sޱ xRMg #tWRK;k]%͎1.& 6}}Ӭ^4kv@e2rW0!֐`62,sLxWjJj(BAE/8RH}b"<}'V0: )g;SA0ko>$;_J 7ĸyP;qaKOu%ߵ'F4q<-T@<<Ly6dl<RS%[tx*kÎ mG 9_{VC:J~V`I7Yf^w=Z-HF.l뵺tQvfBCh9snD,wkj 8`|zVEa/Z ;4nܝƦ[x|޹YעjY2~:@E7-0q8*A8˻{8 7lWz/~94a:nu.:W..\=¦VT=gpP5qX]ʺ.\d?Ae=k4SqG^>w/~~}r -r/W۫/.V/W/V\_^<hnz?|~s~_._ u~kLtlot-חFo|eew_4M6Tl^#~~O+(a踺k~9KLoL~6٬NmnIО{?zzÔ3DeJ zWtr+Q*% k:GGcF~]Wpߟf޴ ER5zk *gMza&n4ԫ򿾞;:??RW?Bit0xapqfߝw$~`*uRm  ~guY=?o6[^qٵ=3v[UkxyS?_E>!7䏧;\d0ѐZ.'z4OߧGV: ٘4dSJ7p2=SBie `ZS݆_W2(UHRm< EcTC|>T@?׊2c$ďܫ$ Q+Ij,~rcCӡPևbB{Mc= pJFB8 M~q3~7:ҳZͦR}zo*Xn23|Q5zZ!rdi>؜Hq(m(i1~q|;_T쓋{q ©nzގIP4]> 3}f'I2 ˄0E2NFL`KB:,g+L,BW`G1@΂ O}k'Ml 9k8lɺTL=g8 q|FF|O͛9WoW T]\UM]qMIi1/M̶1 60z' ZzU[]Z'8/3X4ǔ2o~BRI8ʕ.6=7 $thl_J9~;IusikXg*iqǫ*j2[N@7&夐mF.ZW3 GDWbr"xwmmh3gi~L|s'tz:^m%ݸ%^lɅj`J0]: Bq_j xqu&Tz.I庑 ;1m-*됊+u^=\`J XSַ'aݹ`r* KT,b2ӯ,$Q(].ݤTKeZܓ]u[8/q`) `ye6a~  eZU$pU0F顱x:R\γ9l^@ el#B/tP y,f,ch)JPXR#^ϙb7JmDng=v67nmdҎw'PZ1:=.,f?l2.F9gLMߊseQ&+7}o?BY"8(K5xaq2x,9#. quA9AY:_%KFr݄b\wAcEo>>Ȫzfx2њ=Qde w6؉6hf\0ǩo9 ( a ?FcHhrA` ְ%|(+[nLN4eGӱ*Ne0ے;ꉓy| tú i-i:ԩk:CwA娓~#i ,tѐzeuY4c,wVFSf H>!MQ]ГI 3xD; ?E7|{W}[zLգq~7Aq;_4f:ƿ߿)\~zw}p>tG/Ñx׫p^|y^pܧؿn: .JIS2&7>&5Y}[5Yt9&n+Inx&"B#;XDC":Oa0Dr$#d(;:#͖L[`Ϻ2!0*dWk{91 qxfa\\ĴC&IըR0InQEx//';PБU#<;4GqxQuH.0qWc@)ٽ(2g3p|:g pO}OQQDcH%jNO^qCfi(

ʎP)"QF>|hg"r`0|!K9<&$DL@D1S,#́z,(@K!TFH_A6ϭڵQ9&hMOڅZ8.PC'!q/ d!^0 (}qQe]L%czKQ]avjT v'm(ڡ6z ((ԃ T V#^CPLh7#"ApDVmkWrN>V #!…)tU{>"D.Q@.c "dT2PYJ6YJ5XI}0v]gKMn΁lV rYM,5FKFdYpr0aoS +zϿշ翇(:zIhDs7nTl=#<A%H+A!d;z1kRLL_6[\Y훳/[YֵZXJD{۪M[iSr E/67UdhB.ioN7&W_X$D'\J玘uĤ㎘t9$t8.xYf*䔳Srf/*1hQly|KMܬx]1Q~G6l ޹Q!c;%/רnaǦJ0K7 njT.>ѭg^C.aN Nq봼r+v}.qZ(qLgh,^@d!,pY.MXDǤVQk8ج=v; `X.[f4I'럁mLU4j|\iUpX}kժ ؀FW(H&  fS@N#hZ5VZui;met4(?ԨpώF+z89Y2=iGiEI3Fbbcw&nCu#ޏQnѾ;D]E0l1Ku#YY,#Q:B[}XFYGACƪqJQs4}IL|zK $-25LGHҲg%~T1D{|uQq&0٬"g)9XCu{l'3sZN2#'_*}'+k{u!oƯ~JފƧ¶:eE7 ="{)),LGV~{;0N\ib1=*&p;vN Na SH$t-1bOv"~5o L&Ii&8c4X*&ժ3~曲{l@plD8;Vk/xC&ڬ>#w2VcC 'v5D؍wp8VjHXb* wUupZJ(pڲQ*ͪR 9gI2v/Ep}Ϣ{?,ԫAiY']Eo:'s%}}tWOx^ D|NئPxxr-p+Urr8_OW_u2H8Q5/yAR^>%|_|0d `E {D 17-1uͷ-!-&E@Fq!'P459)B`GAO|5!_m gDM @.O›;j.+@VPMR?F=9yq2 'uf`m$ZȏpF+M$Ndޮ3:?7* wnE_@EnBfUuYtH*՘hO"( Q=*萊3S ,}Fe(žFIj!r{JiSR͚=P.\ u9#LԮH]&RR{&$磱\ [`Cܴl`2BGsp,%~.(nAOǃ}Y''0ͻ[JfCl]g.3#L&6ZhHEpu(ŋ7?/Q;גO*SbK^5Mf Zұ'IDl2ԑQ>h^FvҪ~t댩]Ghin1(1'[0FW#uvHL5nsQr_1S{Kqe %bĉ5qukq/XuH7o-o<;L(T0{~LA9"{XpSGeə0ii[츾| Ot!MxzWM4Nin--0Q:":*rv^rΨDZDR'G'J@| %5%J4z톛d{]ht -8Z )`,}4MEGJ{fQ?$U6~p2Imt'SJ?V2,6(bp%Fſ3}Y=o防`cTHr)NV=m["o΁c9<ORO9SWVulOl<ï>wKmwLзZJWt$9z庼Aog[A1 (2dHK.` :aêWk7d,L٧T! UU]N߃,73bUv6u_9 (UY\ڳͿr@(lDq/*F{Zt1:׈sI!$ARjA \@9/yq>":'9/=A@1y 0t H<ݢAT->Krq\4~`: P\',xB :X޿V8'ue RB%A&n&KILS;+9,Q!CHcT0wd$ #s?b0.)@\3 A'IN h.( 4/dl ڐ,󚾨kqQzS o|n'ɇn[<8(;܉)R/嫣82w5<S:?>ژ!Z=+N>a=&1Ż#Q6ژZ#pEgB@eD V|mLy/D:ZaO7*gOaF1{hF1{hTzl`hTDġf5˃-T{@pRPʱ &).tzዚ|e}u̇*-6l5ɫ?+c0]7Ͻ쟯BNY 9e1ŐSV 9}W>\-oo^&9gM4gi !;5k,zY{M>!`2~e%)KAN`s!\E6+rD(\\JKGh*Ƙ zBj dkjL7Bdσxc%nrEx͍h/(Rm2C&3kBsx_܌&J= 'q"B j8jU WLZx%G^^imx]vA|e$ar^˦1z dA:r-6@'2c@YAጞ5X? d{G[~݈42nD̡+d Wns ▃hY ~ (MhSt }.d\tDA—R$Ʌlz7%RXmuP7(^~MQlQaCT PͩUG!A~ɡЊ*Z|"`dşO׉!)9Ϛ*ۖ6 IĐF{5V)HEPxq[5$Ʃ"haՆ{x% vm/RNS.P y,RTy.Ā;d h80 Aj mTLUkٲe ;[mSb W[bO> c-E69aHnT3̀=i:*"~wJ6)-o^]J}Q/[ӕg`rh@Qd"gcr.&'deUFENHs֏81\([DNVp&JWLoJja*նj!#rP*(هZJLiR3Mne[%._.UVmQ6H_ޗ{zϿ(KoWMAJ;b}G+\3e>㙍uͰG7K*Զ n3w|ZD̂jDՕLݬt7 Uuz-etH6k_+_)=,vӟ6"StCKQp ~.>|R "ZրN0ShS(S Y{f%^o/#~K;LH+3ʘ`p\kiACص Q NȊع17(E:C{wnᇮlo7'] >|7<7X\ч0DV>}qW;@y:3V.25+/l>2#>Q+- Y zo+P;xGׂx]7i܋.vعf~&QЖnrTIs;^}b;eO#ɋSd!-+E*lR'Ë́t"hzȈ2Su=ibpXT >\9hL yYiCe ;qoҕ"RmdбsGUAX!x儶;*MUx9J |:Cdk On󋱋1&MBRʼn$ M єwl[[*^|wZP}|~Tia>.#y7rZ\晴Ffό I'dRw;ۑ/%P3} Ce?;M% @bwR>" hS&>{H2*Sf3"b#ZSQJ\BXeΙ(ՀRAzk\V) I릳{rRಥ\rr}Ĵc4t8[ߟ|wJst NmN)^w`n䦄2br܄m )uӥ3I T,&-2ra9(E*W2wX&q&u*r9li{qؒȊMf*f:dոcz)i/mx7O,t6&m f!jcot5ltj~;Ol*lSSM8LbbTW LP]') 5[ {YއDECC=c%Q=k =!T띜%Gѩޣ$3?Lh}9SVz)0x AIrޗ400T+}95XvH׳H(}EJ FtڟPc琭0C(g>_bΠeī?ޞ/h$"UUA(}qSvфHe?TR-Ӯ89)J:1AO$UmeU7wmHr(=4hڻ/W|X-;4=d;kj_c!!gFK!e|ƦY-U1CWd[tC = ZB0PmJJ5_}G)u7:rʺL!90שb -l7F]vI;ޙ(e8&C.nx_^^g`.F YV~dGKeZRKIW?HG8[ɩ3g9NEzrZjl8;94;\YyE)1#(T 6 n'h]ˏt~7 IL m| H̢ihU"C;ކ6dd " 57w<L%Xkig̞]{vAGۡ'[;罈\:zGJ5ԕ(TĽ&X1зERDkNR;RV'Ƌb~۝Nћ*\w^⯔&MGJ{2 NsیpSӗ8"(- Y>}J^ 6ل8֦s6wo}=\liU4Ϟ |CtO Qp>/(5L֧3F= R+/DRgeOq ʕ}9{w;L/whfWDJf {}~'jm:"xjŁӎ`VRȵU[[OB`*xAHysIS?2F|bwO'2kgFrJLLs'G^VVISUz'ph(ʵ#Թb p"3P̔AZql%ߤ] ?Z94V)p輪3T\g 3ƌjccdzن6](Dsc'*T!33Gq@:s7BHf%;A/!Ɛ-WvbJ}WAȎHeAv&ffCK.5UݎgsKv ,e~nhơa;X%L^ Мw9ܛ=TFk#ƟJ2u|6f<)Q$Ty *sJC& gIJ$={SEchvUyOFFq}ze15L[mwrz4E+g/>bKl Ddz.MkޕWciZxy}oDleFo!הyMm _4^Fin"?ݡhm gĥ {\ ~%Xbhto|? WwX N' &F֊tXhq#SwcÝ3D5jb9H;'p|BQF0ws`58^;iv3o¢tuLGozh]nhԈjkӍUB^zO9WppF3L&pkZ~m\V\E[H[yܭ1.|`kJ(aEt᪫+ /16XtR 5~Cha +do)T6j,8eOR޹O!d ڗF=iaPjS9U-$`ixU$' 8\֚ >K@眪(,+O` Dv5N!eTIjv<']yPiIbV#H<-n_CJPޅ[ m|Tiv UB3QC`x_KQeK.c"B16Y۷?xdz]ljJ6ӯ7~;|/Bt@Xsz GJX $~hD+*B6ϑ+XE)Hi.&#Jrrpm 3MUIH^i `+<' |<@of `4N_KgZp$P\!U7z GW$&#TC@: ҜrED/3qVdr ]s%d[1T@[ ٹtgʦx=ix1oNmƩ1tgmϬ(8.vLEFl-U@!}}ꟁԬ/2Nw-(iOp, k9ywPt5'͏ov_,R&:FysSGu'sfq&s)0"\ՀWM<giC>JqZG@Ei͂אg.:? [ #m^Kڥ'A4Ҫv? ZK6-Hl?PH-ZBD|)9^* ?k-K sg0hj}R4Jr |>/:g %fJEɍݷq KӔhdgRm;#dhGg(ElEq%{\ ecHuHQ"YIz![I6<>N0Œ+1ratKXD=z5ΒjTgKV4(bb BQӶ%1Gj!W6 ;bp|37 hK|o"l'T}*d>2}_,dn2FqJ:NAI)f@}Ι5yr˄Ndԟws^LPqx,fPwA.ݡ" 8o f6Xhhv9ڍ? RWⷨDo'E5;瑨'~ojYCsId$ ĕ$9h&ė5~kT ] ~kXkHɎ h23\呡Eh.q-*CҹDuK fVK: LChiƘ+/eW:f -FsB XIn96q&nd.Zyc%P9r YdfındP[KiIr:7^-Ɉ.:\`ʒ$-I&:>S9\^ !;&6.Υ'7{'Ӄp_H9~/4#D?}yP,D5$?etۗCF:_02;?_y5uk2=j4;삢o_2*.%h, Wv:Z} 0[Iqz:ߟc|Wغa{SCZyCq>.T})g 7jߍDiO6ҩ.ʋЦbT)&D w%BHnr/6H[&oY4ў}x74dƥ <>XԖMeZXn~ '9Rlq|CnQCʫAdׁMş0F~r_'eI6dO=^.>+%;w V=~R. [xVM%[~#Z 8q-(l`C |{g ZNYYitajI{Dvj*-lB ;)aZ%`F*xˑM'%Yhd+h l%TEsCW9ީܻHeOn3Ŏmf6n4,nZM5Ifel右̗J=3ńMSrb?Č4 ݌qcz1mowL6qzkaJMYe@F!TkU*.[-LD;p-ws] =xYMb\ ^}<᷃۾٫;|rFDی27RX.'%*t[cUndpݳӸNp)DOA&"3:V5)FqqGvx0 tZBG>Kjnp:v: ?Vh8`#F :~~E>O7q,={4`͇%˙v'+<8XVqFr>Ɗt(/usGW+@D RʞtVK/$h m|F5МSee ,gsmb g5*$eX]ʴq-pMtrY>uQǼB,D*99ø%a0A6ZK!>Fr:ւ*F7kMyWAWH>¢Y:9^'XTGj](xw|Gƒhʏ7M 8X͕yJ|-s &Q(!-oDA2Bvʕ{6# /0T? ܇$bM$0)1%-Ey9_5)QCq(pf(R"lX5쮮wTWpkP_ԣdh 2h7F  s5r`fz_?dy"Rnc>q{xw:q.#@-`t%86M 1ge?+L$\X]LK,.*J(JDcpsHT!b0!* GQcC.G?RDԒ 5vCJ[>N@kuh@k&65 zLTCnH,C*ⶠ@zt \7PH;J]\ X#-6{GM="E0xhYW{eT/jwDo3m^LalAQ \u1gZω.l#PD$`]o՟&s=lG\qLZ!>s$q|X_WQ?Tg{~/X8eU@<_M2 s$Je@ {-94i(N¨2&e\S&h'`ZF77Higٗ*)Qt. M\x{HhR6hO)(|>q왮߭Pȣ;w +{1$byl6㯐;{9B\?-]NѽS &<0wet@~ђe 䰝k.CĩTE=E{"<,E$ADZEC?36>'qw} eX-IL柣G)Cyʵf ߫dvFFi4>rx1mM??\/g~Η3?˙3W@W)FT0ˣaX^ 咁UL"G[d=g?gwZ׍F扛PH#bHEs|)dFwI'̽tϪnL'Cr_&zmysqrN繛p:ѥXC2y ,@q!h=nޕ,⣂j0Md꠹sĘh.QC1L4Zr54=6-jWZRs Rһ5 MH}Xf[UH&%4FHzq%$. {=r{ѼR_D@sѼ,T{|5dB03A` eXaQI)o<&Wk_Hz2QGЮ1`ht$!K&:!Fb2V%dcA@ֲx Dn0(&} #$1ٔY>f#f#{ FIA ۪G<6,J1ss`L94)5p3'gyRJCD @[e}TjDÃ΂ݳ@2T'4;m0'Cߘzk xAm'P{rkms ziՄB9H. u#ax{|x{l$3FL3-$ش(5L blQy}Df[&2.k ~:)S{Sv֋b颼i7(aLϖMPQP^k:P-^_w:,+tXN]sbQ詋AdX}q0(r8lkS͟~3j9+A!Wt.cwéz>|rl,olwE %eJum6"-LJ JI] Fx& ȃvP\@eKQff4hr?'ROpZK2£q~; Y^|/~S<8#&r`| Dh~k2+!S"(`v è6E-JT i,Mj} Dd2]^c"1\1D9k,dA|DZnX3*:8߃J~t *-\T/zׁVaT!Eh(%(Ga/Is/ !$xIJ<^'jaDt6v`mPl [}~/{Z RT>]svW%"Ar~ 23CygcӖNx u1a.Ư!2Qno'WyW<:Y= KRʼnP֕C݀Z_6&WY7Գ"*pVmV.5JрGOGwצryv4!J9`,]5Ӽ=ਇh1]F crAOd(M]NgUأf ذ>tȚ]{>W_~{~w3 )"qܞR2fTɘJr^k; 9uqzy}>:}^ffBUfƻ[`ƼT aqc%t~?}sK]VKg$TaXO,Kvߥi|\(e^ Siҵ>is,L܅i2Ӯ楈6ݑ}E&5$5UBQbaRA]Ė6A҆צ NsݢHB8`") }%ߗ$ nMmMD4{m>:dP9#J)/틳QD"ǘ!4p-bchP~6׋}7\yyz OqMzoij2r~|TϢ<9Q~-`Vk"_jt?Fm [;P}ѯoB>K()1? SO$ !Qp%`8PaLAFc5naN'-7^R&`P3M]GCA{KJ$)TFBHQdmVse"z2vAZ띌2ȠMݑ8w8ኑ :()C #>A]kg,(䴸=ra=qY{HH'zl,9W1GoӢBEW!["MLc0BPPmdB<R 38|Wj'3hÄJ̄W0!.:wnМ`mpEH=LnSM4z|B+SRpE_wE_Shhnk/Cɋj 5S G$F0հf*\E]OV08dyP CP@+EY$tմCdq'⑩P ^ARU!S9RCɱy6nd%{=Om2 {]ܖ@o2H魾{jN\tͅFj7J=K45 7W-/'cA۰Ǜ1m֧!\) ܡcr_]"lpBjಗf($W.;dZNj6.dџ&wmȎΧn3yu`#F8-%j?ƑwL m Iu}vz::kMT9GM!L$܋+D#+'=Y?p߆69mq&ȫiRJ qub "RDHM]{Mfd28aa7Q%E ݷcT5=$z` Hj+8f1黷L<"^8p/;ae}.J8W${ :)RxC1կ.̗\c>WIL {Őo#^Iu]{eCRЊ %c]moG+ؑŀpsA>ܮ,~HBR&iiHΈ= 3OUWWUwUELz?uNdyn8(E0N1AĸH)2 Ԏ+Fa9z: Zk^54u*~Gjmo5|'\L_)/>] A Љ ev(Cow_)8cW|R vp֮,TS[CI& Jh 퓹35B9J-F@[EՆ*& #+ G*QF!u z֞:%b̴T{c7x *y`x`Z$J׋@"̕ "1ENH`aV 6l[O2Ft|L OlN D؆!$A xk#B}z"Q`X g&:KAG'J+XG,B!_i`䩀-!jINæʱÁXp&BD(lC#m -r x,8a}f^`)̼#r3g@`l΁0 {7*WQ @ AxւbEih,p䄁i !F<"Akm|!a$i {dǐqdž* NdQk~r:c*J`B#U+0׍MMy@J|ofEHK!C[e SOD!# wLr}]蚸&&c`-3+yDZbq4N!0w`o>Λ3ʕ-UXl^}pάLOyfo.f?ݾb|)l?,bl݇M56Ļl~׫jU ]xgVWK>R,[حB0ttm9 ѯJ9}l=:/xQ$lZryfQ =â >z(1ڌ#(,sN9_hA wLrP1 dAĮzzgHMi/\ek nͬk;G"Nr N#',P1.,d b51yLF_1.~8_PJAJ|ANYqQт 6 0|}w`Zl:zmNL 1TK('g @EkiX:@bc;jXa ΖT՚E %X`ʾRɖ.=Rc[KmrOecM࿠B)uѸQ<:\˰`G+>}nihB DYi{pV\n*W !_,Si\_6rDooIp?[_ہtѪc{߼`7reݫܾ`_3rT D)B8'P3u\MJuQ['*-3M˧qSHR0lcNN#C|?GBbOϜu!z/"5B=n,ĎV+t ׮ y' }ߺO.?^UZaއ m`2G67|zW;- K)涨g>iEzC j])ˉKlXm S' NGšsR?x9@nyt!'`X2`N1;f3z{"ޥ! Ռܤ8Z[Y7lt9=+¾UFtOĻy18#0t }g^,t9`f3k30QĘHiXDx)Ex=?<<'\' lPSw OSͫKxi#!8JSs?+b[&)8-cX>ܬ6Gjv_|V_V//1 X^ ,Ԑ*+g loi+}u.RxhG@ox!k^/68~`!<,1F7,N<{(0xqP(s}  MYMQw:z[J}) 8%R_!jlg#kd|w3z΃K8b CXqCj˥# /JEW4?A[1L#ʼ&XE [턜'&߸)Zj5 7gkm:!}07YJ(ڋ4Pa}_DJ'W> V?3ͺ@1/s6y9yy!D)3 + $si?K{'?Ob\`SRH̐ 1'ߓHAHkᵰv?m[*PT"݄NhnN>E%xv/#s㈜4e88s1 <|% a^k bvpQQU2Wmnv'WuyaY=qFvz]Ypu4bEpaF@b P\YsqdJiS /<`wu@^ujC8Xl@dTFXaVKn lZ;˕׊1zޙ4KPO7ZV!d)0BpX"*{weȱJĀ^UXi#xJi}0wfy "9COO&Kpz:%6wJfDI-\uknR!O-g_-GNSlR4OWno5Ix9o w; H({?+z5Q}ʫalq'%:*;t3ztV6SJijlړ(Ǝ&"FV1FєKx8P.,T2>-`LoII9CSU.wD0r\iP8-cР3@;ի]Ό΢IG疒3L(h' Cb"{vvVvPp= J9(&\ (7v$)ؿZ)@Qm]j9KQ씪 c2[v}?V&91 &p}sw 8,*櫅 l+Nyp ypsp;M^؛m4“n;p9?PsG*bXԦ4q=!&E.R!;CcU $}7i3)[3 & f'g ɎM7e%%@kY5B_nSBݦ}?, AC4!$D.ƒpiLTq3ZŠ2~ԪƏo`yMwdkUBv<055kڎ}>+.&xsfhq]z~*bɏ}t|]+2yE$gxKpr40?{ܸ/dweҸMRI*u6gL%) ʒW? R;)ek*t8BB{L 2|- Xets8viBE};`ze'װ+Ԧc'h"h}=:ZvFYZnJ1'yqŵ%=t|K[`&/:G饓]oͦggR|KH A {!RR&I۵+Eu Moij-Zܽ݋ϟMs,gT#R$Yq~sZkCˣ(%*A8%Pc![qJ]9S ;~J  ܧ9PR\x"maKsDKku^be21Ij'ǂE  X,1.11(J/߫M;`Sx%<{yh<h?,mK0ܤp⸋Y09x * D i*biSM8C`;A2?_5٠U9dM&z#*8FJZpAk+EJ(T \J7cf8eB;ɵ&qqb~"{2~BSC FBT 5 fWߧx_vJ{ y^sIWJ"4R7\ @ gkʼ6;[dn(BZgU%<]0E5R X= NQߗ2B h[⺔66q\JE]gwaABUƮi&s=8=8cD xsi~:5khQ8̖VD7\BX4wڙǪ"X`da^ &+C}ZX^su9śsA`'̮p.`|~f3<ބ57ACwdlD@n6q.5~J!/OaF/)E@hV3ħ5V.. +}?}/tlxIL%i(=uQ)׺rt/_-ԊQUfx]9;}8N[7w[[)*NIm.,ͮr9N{M+ ;N 7=Bw+݉`5m|hg,VM Oc>_f4P{nr ^)90b!_URS֕ӯ50;pH^k{:$PB F|x sE!*V Gl:н{A]TY]cIgZӭsKkq n h(Fl8[7)Bzۻu1x8N ѐ4v@ڊJ)\WTAc3/iN"ݠ[Xƍ)7eȵY1 .\Ϯ{gsdhgIrsa4Znvyv0?ų/?H~dz#7:l|?ݺ׋MϨԫQ)j\aF]yuEas$jU!tK˳yn7Aex gUl`qD<,NZD0Z11RG© 'IdSlMEeǝ S bsјY(%DQ űaRA",vPMbe7RѕF] vKM-mQ_?g8!Rn# غ.( Ȃp͒9U$cSӯƠ  M Mb<"*MRV"R9 ]t96tMr.UaYj wmG0YY4{L07n2<[_w<;ӧaẎ"I!25~`U6r}'c7>ˆf:댻GBy&'_@&+ad0{8ا?&f;8!gszx([OnJ~u*c!oDlJqA mw;`%)\۴w hYVqߦCAn;A}XDcɻhݺIݳAO~ {4_|_F}4~>?c7mM~"eL7Ɠ2Io<+ű֎Z+$é1l"-Y)q Id(DNs%ȟ8o/л|Kg5/ny| 32ý|뙹2٣-]E %}AH [MITa$$YR`RĄQ1b-RL0ܲ[`et9]x'_=_?\~AS7\] <' j1i"&2J̥8=Ku) }Mccf&t$r Iµ8*Gʜ*CR^H1Cn)zTtݞE%P _(,0B+JfAHZ8KGAItJME.HD. -!|e1+) Ŕ3$ v"#%_\ۥd"[.{wq gޝOYo.9 W% x?^ nMgu˳J)d\'6(?Ԧ|M)@Ec`A)΢"XJ451)G, Mqb4-(ZЩBT^f ,nwf7hF m_KW>W#^YHʮ<cLMG=?}zz9foz_#&f>=d~ K=Be3Fe3Ch{Ɇ6Vi){wLs FQDV?}3L^*qh TdVknfM6d 3T3Ez=Y˪㚗*srQ~M9ɑ{MƣH0KC5ta2k~??ougSO/Җ9КTw@0 (C~20ѕ(_f'Q~rӇ,?}`8~?߻Y3(39F^d {3{[u>rvFKҲ$|HZ)Xazc4mҖ㓬N?bqT +=z`."CX01x'֪#v{\E@Ö Z;u!JRtv秿K= ]*M 6*jU6=P$=.Ti.U^hPZ"ѨnT[WSõpdT# h4T [d{zk-  7Ra lm *TVR5 Edܵ-äPGX`%bTF)wT1sJrYf<"q6U0IZYy~peZjhrubulBeN^Dl-x:o/UなjP9!w*zD e,.V(G0B8=?Y^\( /y3K|=[S'K(@BYyq:;`kĿb_W: =J.&fB/E| 2o,Aoc@_>b2ohPMA@ wOZEuʗ &GVӔ2 Fh SSYShSB ̓Fj*8*Q74fԧdlNNZm,{~fsE|7iRڮS u[Ι5yA6;; 2 H+[)[zƊ@~InXJΉ&@ -+/$hREX FcP.cs- dIk|("P #!*+UВL"L(}HM7ܡV$" |ϲ`1LPs‰.f1# q:"|M *K4MEU,mE :Ya`8C]Yoɑ+  Oyg=zX-y!TwSY}g]P 6*3Ȩ8NgAticо7j_!Be ʤJ+ r`Ž +2>cCC;GADlBu 5γ3ǀ!W Δ^! ~ X6n;/7e]qm\յ9p=> ۊW r)Quodw]\󼛌:V(U)acYa+p+wyO_goϖ}Ζޞ=_{};q>B"M_~ں1"oS(ZccB 8{,ǶOǶX$ޣWLg&gGӵj*T}w'ê>NWO wRXpVPNo;>[fM %7Bth$ΙR;.<"1ݭשDhs6`2< T^g da68u$/zfyKoh.pDәJ.D]&LuY`SF$"0 &XJ $9FL0 f>Ԟ7^ 5â7Cl5y?,tϮfcxbKc%9M % `/*!,v'rMP&x` ǀJ,r8ct6CKh)RY 4 Jhp22FoIn3QL2dy=. #<_YYΩXFV߿'tt>ĻP7wQ껇gQ RN{&R8x{u۳-.?#e{Kl>^E[L|{t9(b7%{'w ʐ a,aų6E-'cJ:aOQ *޾"eQ cM<,pbJn0IĎb01 Aܦ8-|Θ@LR}bzy]򑽚9)tqiĿVY}g?=o")k^ jGT)#.uT/Fl9W?k5yQlGoF)YčƏї}3K~#W>/G{W].did &8bs_ƓdXBQeH^Fi4!~:[ 1sꦈ;(!$<,@8k`ъ/|0Z [ؿKfu,X1V$bHbya+mfzZduLfCW_f$Bƶk(rˣ<PĄwr1hˉVX"BSM$j{Vn[c# BȻ>;&4Z:VX3`Z!-2FXiLZ`HCNT!mN7j4:\&lK?K`΍{uD& (!}ʵ)Gds D թ7QG'.>mSRegsv|8ЗX< U@O,+>5b)%OOn'8 e%͟r'/$TZVO',&@Ʃ Wd 4Y0T97 mOV M#B(vg:^S& !,,P(@y5o=Ou|we%g/SQ^@S,V dž՛N#u6;xgIT~H-by:N_*fTAi{}!8b1$tD֥{wг:YtX,PYOV5P썇Ğu9֯a9?jW\D$T͓8I{"TR Ss"Up@v0Ψc5V{"m) O[ l[CXzc7 qqɀ p)Pbm!ə$p,58'-֩K\]CJy0r`#l-H! @RLZHS#(sx#E%fPQ5!o5Z8V _Ʋw"NE_PgHttwsJЙ*DOŷw<KK<室??vJ^m=Hrv~n@ 46WrtAmbۭBeHwB")(zI`Dli.Y=N\rT)7EYeɃM'팷~NTvx2KiWO1>{YZus,(O zwc?K<-5p]U7^Тs"ў=F%\_Vnyy YaJ\T hIB~"D8Vni7YrBi:Gv:?{)ƪ7]vkCB~"!S"S9kOO-Q( zy0_m) ȩ-xЉ3@wv*H:mƧ0,1hEwg:`qa}=ΊҔC+qnWu2ΰF!WOCo#}}Hun#7bwen^ienv7?@Fn`뚲$.~9Vr,lU߆̮lw^&V8 cTel5q?BXj,*,Ĉ4oڭq]1j1-Ekw .k rNt\^V]CQƖ=;:0qDʿ\p`ыL$6:z|'Ƹpn*S+g7ޘԋ,ǜ.s5X+yPz9՟8߷AHiuIlD"ZloonwqۍʹC!PhU,SSkg@ow~X3depFx@sXToFEMOQWEa`;#UJa6.%Vp qA -4No<s@-^Ybu MKԩ"Q tH@ M6^)âb=8m@jzi CDf*^` fe&ulfāƄ8RQ19˰"(1 ][o9+_3-~1`C,v$ݗlmkV]g߷ؒ(EXAVX*VE/D5#WA+'VIB;RxR"_\+TBbjYum0lWGf!!\DȔ²ƊvDw*1Sum)'iMvCB^,S qdM7`U*/V( ,,//TvL+\q/ENpSA f"dAd@QH] 4+Tv+p}g>LoI51#u;m]I4Qda=ggB6k*Ӓjdv)vQԯ2c(6qXL.3bK )n&ZD2aMa^.x[&uZrtkoBYPeS=X+GbZj~r@")jn?{T(P>*Q{A"+m_c9 n #E{+BD ~OP]4[AљObN D3-53^`R樠0İ 9BVp^cx2s1msئ") xhjJ!1[))MHQyZLEb)eq/;njPVR ^'ET;g Aęu&c>) RQ1[8#O0~¿' f:=Yf.~ `lr Y?'bK>E`??Hf-7Er%t>-r~񋟁s5LNO1 F_stLұS3]{ӿYȄu3Moe|70 C| ( H3$y X2HnU3XzYhg:Sa=#V G֨JI1R xF T-s畩S&2v|i|#lk`KCr]>sg.Q}wsB r 5` };37a'"A$)~Dۧ7'`O |oo;C34c3^"a>u\Q{?bIHB ffEAl "ðִQy kdPJza8xMҲ2$z"}$) T$2PV 0+C}1$(P2 Ӕ0„m-dT3τ8!cQfh`;)ɴ%VMv'$lIE -^G+oo Lf[ MY/t-e$1hu*HΤk$j&Ofɂ 魭DtH-B[^4<lULx;5QVX nIIC Jiv^p3ڣFCEL$6&Dٷ &q,){ !/(qb--pcZB[yX{y0κ?BR9PSsa3_6b=|7(EH>``- +ҙji ~֘Rq+l^䕻[qEێ7Ĉ53$]QS+㕺p۠X&Qru^_úf؏&϶L`ɏPHIR3Gk0QWj_/cH(+VhLtBI:&կv*^'s[\hxCbcgl?oozc}lp/Ɋ*nY>8sR,pvz5hijpN2qH3Tr+SJR.c'R̘w)AJë/>sn.üq*ArIQR$%lzp-oU*Q-ƳA6&np,ܦcI5@RD-eN2i8g>Cx nGb$&5Ʒ&]PQ+[61](nf-{wu곾DyCT0Ît% #m~=kW_;z1xDvw[)N#dynO w܃N$l,#%{ekḭӨ_sZT0~`oJ8L1DDYۂw4%${B7.s؟Bg!Jn| !D3[E!YDP.m`oRB$XnS7/j92V)ߝpe+wxuQ2" pR9pT;e|MkOZkc&*W5k$-hD@~ԕy4r"!8Qk"90e 5TKis CV&e*1RsAPb&˧)_^B گZg 0{y^J^YQ@h b/ o V 4|b^6qAciKZXF|"EDb^{Txڒ^P)t#1FF0=Ub֩TMFQvEڧ`th҄I%F 2ʨ6ְdbe)zYne[.10, Y1n!-eb.| {ccKJ5`F;xZƿo)w(ì(@ℒT6̉~R}h`*gtu)$M&)Ij,Sd)^UI7)ByuG{m2A35KQ+d|9iJ9Flc8°Pq^b qvՒ`*X )$bTٙ[V8Ra*BIpTԢT(oNZx)'X!LtƆ=UQ8j:s\8CZD{6]rʂ)<άfu͐rm'>xx(Iɹ I (6ʕ|n(HT_'݀K%g+0VVUHt_HZ܌s}0zR!W9?NWOeƸS6M\Mg<V# ILmHj#l`VIaD|pv~(6f`S קg })3ӫtl&gW`t>Y ̰77A53?`ԯƳBRM ^b}k-S)g:a׺Ju|5޼I].q@.k on: "z]eD=bF4q#קv]&Cw{0~޽."j -2t1s;^rUzCS7Sh}d Py7㍷ F7GFAx=CwQFF~8g5[ drժu,ⓇZgޓ_2V1.j|؝`7ﭠwB<& qԱL9wQuitcHGn)a'1Uh6>r]\ pb+#5Ś+DwY`$Ve:xqVd>jĞ(ih)i+uXzqGKO 2Oq܀7BMpw GDofvw56*Zq$ IpkB^{|u8>fdҏ2.nh?Ȩ"d;.e筳~▣)LYRfyZg!8K,|~U[g m3SJ;żRyumR0&8U^'Aθ ޯʏ"&tDxaLDHX2ι ˔!R+uQE.CmS{؞u00 24h.K`+ɴH2g0(4# 8V602U$jLTo+yCfSo~$jIliq38i)/35ED!%.5IL8R[ɱi\vvETGFxkf1y9\+0>&`7(ޥS?XA| s_uDi> o??lCCEς7 |ipaHI^ޏ~椗 &w?u}B ͚9{+_p*O`n~*7ޝj.ޑԙ 2f8&{1ziɳq* {?I^&z'J}8B71y Nn]#Rn{6oHE("V>(*e"HQ@ƚK+L5qџQvq <7B.~~&PzL04EemdMEJut(Iט|+ j^ 0;݈Hԟ]UUw~cOl6tuy8 N}cIm$%!sF$},:@>Hw[XbX_-8"v#t}+Qo:蠶uܚvtZi\\&(g-|ZLfW#g~ g 7c˥r =>x`?ɖ!` N2ݧPI$] ,_Z9P)U8T`_YIiEPD^GD55 <N? 'RYT$:\SwɔuJx@}.FDS曵(iFGDzSYp=\ňfvnw#6%Z(i83,93&3U/Սx CZlm:p"& nE·6IL u\*/*(} KYP^J d-![M~?f:V,nW%De x0t_wS4>/%w$tLx_`s<(:6fh(~ Ug:0=snT1}QW-1KygγZ#Rۋ`sfr}{s^9k9p~w1ӣ/bjO>IجT廋&4yyOM> }L>^;1шӉí[\Os͘2G)ÿk3>mh $)o[DI!xwߦq^-g1r9g J{$ęc+R6u54mp*. Ԅ%K>JyUʃRt̃5 D kM97Q΃d!grjɭTR9kNQkoy_s36_ l*槗 1Vcg2u:YPY]0QU;ؚ,R`F/v,8 \@)#%LөrTƉ![$80 ӫL r@&SNI,㰂ӎ̘"49RP ɰz;AيIe%T%*Fq` JyԣD.y!DaE h0˰#%cIpx 9'A YXQ J=H1&1T"^*EU)JEU3h AT(KEqÂ(a[5z”,l`G&d?VAfΆR EܣuWUc'S P-J!j*=%b X!K:'(,"${J45ldirFd "͵„M"oÙyC`~`4'H l.h *B#A8?<ҷ%R:=[*PI2(mEr@ !ri9$b ;#i$p0k<8P*@Y:DhaБi AkPp2ZKO T("PĠ\Xa?V!EliB8r,ˆ"f(D$,# 5j8p\Hcp$ČNXJo7S$ (.܀L]L0.h9=]Ͱ -?uh,j@˫\f\oЛwZ*?ɉ FG51GBi$T8yĖgbVE3 {Hl# U|?dj5ia{*"(EǶ*0YpbR.ID&'$aw#I8`d*.֔Ӯ±H$z|y[a,,뭢#IXvR1,'^YQl{&p dt DJ78Kg),5 I#0p"H(dU0 qP QJ5 9$E.8tYҗ,#w-  \a8[ !%9CUoJQI59#n(qg 6x:X@;b]-^]VpԈ~n|TE4#Ի $5GC.ׅ8a"\.LtA>pFpϓ?!hn iSCo0bxNT6uArFHlhR~z-lfpQ7, htK샥{C=RLiy\D) GinE"z -BhRf\~qrb6qJCNSP!/'k#(3c͓QMÔmkG؛Xs%)n<J ݷ5 F2'ZdN[OOy/`G~[ăጞ+nа4\t2?+'ڮTg4R0)#2/]VYKs4@礳<t@ /z=*O\'_d 97p. [9?ǿ5x[7G<ΝX޻Y¼% 4 u?@,CmcљNS}@W ]+n6GGװd샣5v֬L"+N)HP:4|zV%k1ՅV&W0ٍq>'{fKHkޝ6L7wӣlT &Zs1?ћLGԞ+ ͩ-,jNK@ o Eb'g=͓S=ŚU&@*  ԉX%S\5䢝T듗-.KyzN̥âYi!]ѭ _.|0sGE`BzgTƠrpOHvtfBBqLO]ۑ|Rs{za = {8[|/fVҟfwYtv z@WD&0b #$ֵ+dr U0<(#ʹR \k&{ZF 4Ȣl0I blȝ#\-m1(k!ޫ4?q%OǗ<|y%?f 9@x2;lKl7k2]X(8Rfu_dX >_o`NCz0 =7D*Loӌj;•2 W^9i (.,'h'z@SMFS"}J$ -8d>ĘǬ]rlٌbh`C Fe_wKz|˜^EXsw3Y,,̫daf!xY][j}'ň{o Ǎ}x&w]f#ºGw~@j؍xv㙓~+bBeSL޽}a I8&3Ū"I_޽{ ;WϗTJ?Tw7wg `^_p MXU+`kfa.J8 lD8 K@n6~x:vXW,c*e^'.+: nE)Tʨ=M1A}cۖW"Q{P uEexcE1Z3q':6|Eq]ȨÚA /L[cVCpsj'v(:AGi<{FYNqn!WhCli`\fvAnޡBk6]yπubq__'rrmU䨕2sSB`DF}Pn01/:5qS|=/ Vyۯ+jX*Vϥ 1ޚv%|/EtSZ!m'<~1XU k+DRȍ|<` 3)1+T :WT4EAmOxx |u.:pl;+Q\`v+;L& Ī;\\,xnly[z&-׍ +oŶ{:h=ܴz3 D#*/n % ѼgUYk4=pX-%텖JQ.^O b5wO+ uC)iE_eO7hRIzib_+fk?2yd,YɳL.Zb$2bڈHG, oȇfu5gl"TGØEhpE82T͹,T`c8:e[Eb5Q?,GղؕШ7hmzVqыx褫{ZwI3tk\kr3}߈_]99)b.꩜RTDBdg`?IZQ-UA4N]mWVhVZ5C%b[s$`Ha ǥD⨍&rqas$r)">aT}fo7rz2 ,)j (FピSw7g)};Hq(󳝅/`ur-N~ j=?|Sr% .H?=yG 2)-WM!:Ò\ޜ|@ZMM/{Fl]ھ J(>|{rygL=R1M90_WϓTr;" S)1=\/1<+My Ĵ"}KD1}.m}(Im%V4jsu#)z.Y\x+%53 Xjh9dSt>cެGSW{,h9?.>x2.WaM{`[lkNQ rCRBx_ۏHj. Am+mX/h}0 MvlȒ$xo5)Ke$ELUuu]]Em,^#ܵTな:AE@UySݠbGUb6 **XHK>aaDya6CR &15,R\yd`EHGn_9VLq笃 ~<{Ϛ՞XˀgsDfyG4uop+Ť\nfL=  -dSS۹{0&Sڑ`K5.8Wcru:?p1w.Ӂ|4nRQ,?B5rNk8MBjzo=,o%R8 g *]nn*D[(>:.p:IpSSU!_8­דWB1a1vn}ޮuKptCp )]jV{zE1ZjJs@넎t;QڜnUp΢A>?EiwqmKV#Z˓{הj#DڰNmm۩ɶ\}JT%$y-\rtr\- ||.]EU6‚9*: A&ݓ5Sy鵐0~u_*&gnܗ~fTW3V7E2n$獓e׈T,Z il6j橶2@zs&Mm$G94NBE8eՐL Ҭdt汤[(}]˗}NbM<S Ӣsٹxnr1x=ywQPit@`cLq !uܚJM7Hs340еuuny>%A7{՗ D,5OޯL9Q+&cHB|} (Jw[ _QE'|7qѠ6B0$˻ʢ_|BYOKai~/MM XF,޼ 9, R4+eq5,t߫dRbݣ},Y#wѨ{5t8+h?M\g|.oZĹ.d5%0K 6 l3|{ucҍ׾u8?nϳ . ߬= ݿ`(S1aҎ4aD$q EZI%Oʎv:hG}~0?M=ߟatyq9{ip MF6TX_X]=_)`w-FVn =©RqA邯BQm`Z+KX*u7nKvyQ wvD \rVW%bBUIIXYaQ1'ScbNxKr)*/WR,<XRrVÙ%!1jn6  ^7h*ta֌# ("ug<믏&v\V@l`l1¹'M.dlYKY^Pql-DRiNH>W)UIv LR`+dTy1iar)FZ$ҙ 3DK/ #W:+|FY>+&l =zE c|Ԥ,~(ϜrTHMI}}Z.MKr*~)Gc?tՃDFC>K'D)Bn^ُQbGcs7/JS~>}QL1F82!)A#tfD,FjSn2 IRAF$v~uyc}sٹqs O:]v51 9Rw\?JQ.UK$]¥ѺO'-*ՖqR#<WIGIe.ߓī6ʣu'IWט'Rtݴ`nWMN)E|b>hQÔ8nKL$,Ê'V2wg4&Rj2LscѢ uۅ%0#Grlt碥}n<]p( rMFv λo_̰d69`{pJvoŒ+TlBm c)T@^"[}s 8*"oUnX"z0;'RkQ X$}kpp PrAR8 6S1ՉpN 3ƋYVX m, 2KsO9X= JFq#>mQ\f剝:iVOL'W"SP;,pACWsU*,s{`墰n}0-7A]]d)h"ƄMg%+v^AOj#ĴdT qop?ڡJ㡀 Q7sky^~wwvwaHŮbHoP?=?OU"mFZldX(2v,uK/o5μ>-mQ~~ܡh5Z6Rۧ()BI-AqcOO;ŧB)t4^:Ce[)*uS4nuT\5V_-$)w8b%r$B%8*XXɹd(^3Lq4ZXl$5$R͑z9fmoȍ ,nZ Pe[QBY 0Qӄ *!4f"2KI"G6gk1{xP6×^ܗ>o( +$Ϥ>YX)fͿw$!FAl{[/+nѼ^ 'M{d#j\)[]H|vLmQ|ޫArLk56?އaZ^0& ֕ӨVpH84߬ *h.^+#QcLk%kWKXoB0qMe=3!Nyf; [fFQ=gg`Zu)1E^Yo`|)4l|BmjF+nl94LB,.*5|!dBsj;x@4&T+&XUi.ͤ6DX6sk3J*,~1!`\)kq 9~/LfqSU*m; ]}i8xE0aƷ@ndq7Nw4%_< /:{ZώΏؾ+[2%Ln*n 䅸dw7@~D@|W koA8/۬dlo`@_0{IK͗<&;y<&C d2}}Vjo&7W8w}{aVCU._kn^[Ni/pKm5R.,r֘"=(v˼]6#mB/xPMi.i3AS-KI[F19R0{qB͇>ރ Z T)C9F+e.oX9khA]52lL3l'i4a,n$`! jOs))k J2ih~I&uP%hjan$ښGݚ94NB:\;?c@wL8BTݨ'VpK\+ $6XV~ZWiqN7ZնCC\6Dãp_h,_%>.-w T^bZ_(x{co7 Q[ rv'[emgm[\L k@08}XbTqR{T)0W4r &3hS8Ne)TFYX̉dKӖ " &1Kb*3LjnuiЅQ6IUfݳ$*ÚS4L>!2J I4Dx("II,I!jRf<^2b 0ScLjJ"2 LRBub\|jҘ1P0QAJ0d\1vTM5f5&֪s\[t4A%5G\_1SGD$cR Y)3gZ(K* ɀAz&@92;GpYcaK'R!&pu*.4= |@K.\<ƫ9enljC֌+\S/,AXj'"9Yy9w9 Nv} K =4s 5Pm/)BS*λؕ~a___76ɛIO/7?p F#hO}QT% BP~Bs_%.Y;vD,%9^*Q(ǍA_23Rs_ &/=|*.nt-= 濿*uTJvKsK.+#7T! Ϗo 3u/.bT_c"l:ܑ{rO+wcKe+ٍRFNcLC|E p/6i&`o+{48 Vo&{Oٳ$^M}N2ض*;;Kٱ5[t?4^L4>yaMH/))=Sbt8PbZ@5RU#,Qȗ̉qrsR5A/!q4=R se+PEC.=Om.#'zG&3\0C6I(Z#/t QlMr(9|_o/k) PA\ʧСucC 6 ᨧ5VKUaqZb`M퇮fh 5]Co9yWx$2y|߬X_{$|hBVI_mF&[hA6 2dʲr hvl-f nl mZCQK(R#C 矇ɄnDzI  ҄`KC {aS4v"-?:u6!ʘ%#zĆJk(2r@$RH6km]x[oWj0?e;Szq)ᰇah7BZ* XI2KLSA)5kR+D)A }eC&lba+ȨɥCnU}+Ф> xZ9MYT]OȏLf=:\piHb8TĻP5baƩ#0)ӵa +ּd[[Ho"CFa;'Bk\".VAhHv[iPQLᔡhlZcV\oeCD'$$DL1B%aCLp *Vk9 2xwTÌD22A!AUNW~ ?g3R)lnVkRf|p3h涿, t-|/] f0n 1'o4%{souX{s3>I0B9u9ə0~U耣0:\Z%C irJ._+շRfT [\TB@*APqoGj\ 'ʚ P&ucZRPiL.@ˁ6*k*PQe[Ae9ga$X8cas(<@ .Dv |!c%n} Xiv 3cq-dz};j` qW º\%9/2IN?4(4Cdy?dXaV7(Fvsf+zUư-QGy'1%9LsK|6LJZ~yqH7󖫰QpiӐSO4>' {; 4I..CKipqbH>bH&P=SgJ'㌀Q+jP/]-l֫BZ5}LDdQ8o(3E:?D'1ղkr ޝ4.e!zx߆tt] 7y?yѳÿ{i7OϏw-* =mtrIu3_+ƌ8pF1Lpm5؂!Yp9V`h"7ŜmS`PI[eqV#F [DE;bONn1r(@.͹Z8#wIJu(ܴK%pےmtkt#;ۍ0!j^KN?$DP=~Ң/׎կ-7gp5 vSEAVl*Y{qUV;Ŵ㢄LarP L%^ rK͍aj=Ah\":sfOOY&2RT9[<q*%ሄԀ!f5aЦ=mqPik'3 HUZGU9Ǩ)񢳓rP2jgV!Oh^Q~gB}Mdϐns9 \`T\QƟ{:"+^(LI7i}S{n:O'&g< on{;K{ lƋ{CT-R_Š#xKdTm#z?yY>!_4NnΨݪb3tQFs%3ӴvXV‡cj,yv#&W ׫+ظ#mub/EMU>:/:(TC^~v#qS' ڔꔥBΜ+B1(tU=k4n kA9^5*Gx99`6J77ɳ<>oV L&1\v€G Jz؃:G͊,#bQ|t&u'#^m/avU3FhjFjoB]3{c_W;a9 bf iTS35%!q|n%W }5DDYSbZ(_ D w*kr] 3^T8h.Y,.F. Io  I.% wJ X J3^#J&[=s@32I*..Ai 0J!h8S*)h@Q%f1Yjm*!6.&w!|7 y! B!ռ܃-E~+; 4+PO޿{},{۩~ODc3؟ldlƒO7ٟwe `}X<=A]&(DPvϘ=B`*޿!/8x)@VF9HvY0P@Z[r*XLy&h |X=wjR&LE ,`$RIJ.Ϲb*8Y$#G r +bB(Ơ'vgZXmgd tm?@51R22 87=9ޥ,+(V<)4W֪,`k 料C4N;ԐR 1>IءHd02 !03)YwK BA#SϠUٽ>fii`$$YʑF"u?eRfozF™~8{ ]H1I@$]89iA; G8bp^LGKeBIH EI<T -5) vR#` R.$BYx s^"{e,@ܻ_jGOn}J-U  &Ԙ+վxwu#! !DE-/][1.%JLMd4l?x2IlۊH8 d(DpRJYbc$댲)S`-W¹؄1R][s7+,ɩ ;ÖMn+MΦXͬD)"il*4HJ^p.ٲ-q8h| 4@^)9Wg>uoH x"4G%2`Qj)&嗻OPQ@N@Spch-'4͕i&]@cgp^٪'tݸ=]\قnZu+e>,J0ou2S Ym8T^ޏu:61O FpՓè@T)q&]a* J\0@ <!` sAP1 1{o:u-F!ATuߎbàbC81~[M[ӻ1qJ%CvzB^ {A5UCt@w[01l%ẓ=L=kk _&wp@rA\Kq4/ w_'Qe\T~<^7ɶEb nƓ3{Pv|wvOVШdɦ5z)T筫CJZv^ i % z5hz#C*4LP->{Mt~ J[7޲/w }HݲN_>d0MЯ^d-y?,YUGISh^sl:0ic{ ZvdY?Iiy©5s~4.ƛ'#dT>M'(i$ӑቊb3-@ji+aplYspZPM]PMOJOO05Ř.>ڶ(;r֏q][ʟg˝vk|y?8B)]H47>J前 xڏab+R>sAQZE'J'\Upn-1>{/J;S!@9i~+\iBwq׀ !9tuojs1{6 $_ 0NDnxcpo. 7|/2B8,$BpTH`LXy*dJfR("X˜f3d 2\2Or,R9$9.)RЂIP~C%Fste d>M\ *-)ⴐq8s_dU%M,T")I\FR5s6_UQG}7@+YPo_Xeip>ƺKYPP?az'8|J-qKA0B`'(@%i%Hd %t)2acך-ab).BEsٚ-֓$Zyk]\p\J6vgM{Lvi# 5xV~bvJQ z(Eq(і$*']≔̹*7cCٿ-'Z0dTq/|u-.JB[ePTV+wD )rI]жC_'ڎ-OQFwIp?J<}|]| l4) թ6Z?hA0QᰠQLj4o5pr ZObPjs髙A)\[ňqؘ_R#캋Vϝ&!.orW}]˦be"KH+KNѺ2bi c[bBĩ2R3js*eb*wZO%q* Tx4rvhr^TUb,)/ t.\*ȗ窢oQ G]fAJzvp+_Cpn`9UEM\هYn7l&R)VNk*RC_ I\kN\ӥ֔nvTXjH)wuTӑE,2"P`Ib:yhr!/ۛ~ BZN+7 CAI,*w*ڼ19:̬w}eC $tpMu8  {qBg7D053р LJϋ{W<7BxvlWM_򠴕U:fѮt%$7Ps(2w( QTʈK6U_>% c1(8q?FMg. 4+$Ÿ 7#oFyCzrՌ(3 X,u'"UqLpd%#'ߟ o7 t#,sPŸwճnړ?mf]ѯ8VSp$JZ›\U;k:-=(姣he,11<ƵGD-Ke*%6K5=4R@PC*e@``3zO J͍ƕʸp)uu%%M\]sTS#E )RWhI?hʄ-,Pb0ҀSƒr"%qYJ-Rbw2yfXIK/'מM}/tBK|m^ק}:v$p12Y\O~'"Β9,ll2+dYT( UiLh0|d0J4<zb/Vk?'kD$~Վ$֢NJTthв~:" ɘYb&qd Hc|I/⤓кCB˳p6K|`WFFungDT*\_-Ĕ W(wpKzo"Zq2T ͭoO JƅDmG°.ghٞ2Ӄ@Մ^^I}w$!dn\^ŕ(ޘ3'TBZGwmfm;&uS$Y?GFr"8 09r"֢e'.^?K&zv́kg'32ω PеW1zut{hH-b)P (c8s+6LZJq`jP)#peׇ(Cefc%it_hA5;8rdFb|dVUˌP)UKN7nuf/zb~'Ite:sR cn鋅*:Mf: ԌYPym ~Wr8upr.-% f|# } 4$q%[_8ì4N"\wGn΅xU"@;}zR5DOK Az.C9c WkyAVݨ;Ʃe\Izhj;FIR MR>YjAhĆ8IݥR2H*#!R!JJ[:s:gB^Vlpu}Hp?vW Ls5BI逩W _vLi%qomh,Ktycǜu\," 0ԢL"3TƎ Lr#$݇C:E=Hٍ#v*&g{r6*ǐ_>u?\q"rGR IN>~y^Y*ڼmO$5Wxze. Y'/j4EY;HKӲ:f_9IlO 3g/"ȃXfRmP-H_.voJd7J8-m~06N_˸$gxqN _7\g)X'@Ѵ9028S2k`҄DPcU7*>#l|:hEӁ[#OEgkxy0N+7~8Xq_onV#c5'KX~W߿ r~'>.8w*xWaZCse̗CB(?󟸼G{乼G{|'3E8W55Ib]ނI$:,)h 2 U46>>PM[D (-eFS(ȇ|(ȇdK@MaB{+"e& =Zҹ(D3Li.Yf׬ UU?} arۣHhF;H~-3w$žH!hv-b%yaQ`-!2haD3)RPZX/u MqX$KySJD "=zXIH3ǎOj!%3hTUƋR֢3+2MreT\ZRI:W Tz'-\4emAtC,$a*ЄD]*;2גř*N8Q'hF %$s^)_@ !T.POi#6i#6ǧ/MbጠDrN5&qV8l_X4ՙ#I_6&f@]Y$O8vfQ] Pd7Ѱe eVYY y*EY ܗ-RJ`1iUղB&x7)͕zA̐JbVymtoiPXSXF R0*ʃDrcҢYAC~+gWlz8o(E%\߿jT{t G(-@GSFG/3 24B41VEȏY@rzLj˳n[A/@ooғ bp kГFv|mm7ո C'[;d*$|RqjXA&EQ(DMIAS SB(}y QQhCf,gFsE)dT]!Uɶk۫ `%L9ͭ4_u*L߯sDbwd8߯vwҫ\_=%?#xIƨػ_T.Xx3W?|~`)aӧ_ 37hno N'.sk־A{`}|D!ڐr*Ӑ7wO\t) 7R}np$IaqqΉXc,H*$ӍD{s07^n"ggLa>߾mz߿\W<6~jFE [kჿFV\haz /'+cf!H>f6DZ mP m9+<BHYd*F_e0PQh0$ @0è8n RU["S*N"`%0][,Br{^C9wٗ'Npn3vɛyoztgKimHoH~{Μ!趿BC_'[P^ݿg#ap܊%l\\Nv_~ç꣚!^m7g{WfGqB#k aեl0q7|9鼤+3~KЁ% _+\v0oPT %\9zebYW}s [ cnahŭ]/]z04 1} 㗭IéK/sOL %`2GA HJJ. - -qu]ZB ͩ /@@A1,ٌyFׅВ\Z]W}DKZ]1hVeW;&װkp4Ք~٢y<՜k>n6 e*tRnj-SaݧDR6/YNR}2\2;Z<|}vp[+M:!w#CخXg:R{ۅZ%p0РRְp7W&(摏Qq*rÔ 6XuuF ׼cԐ9P>pm86:1 |~p=ƃm8$x9u;$:9zنTAG£ɴnpEyG?^g/v^yF/5;p`=Ac|ɖ'[9]/ WLƤ`W҉]4-ڽUU[:+EP&NK!S~}MD##!~ӗBs1-,90Bν͒Gל+}t*VRj:P0 R%@4@+JG%!t79 "-S'Z`QBq '"\)=g` j ,FyS;ӠKj 3?O2+/٣H9#P%]Zjrgxma=:skr7ɝ bV[͠r'"Y "p<#aZOv T)L8S@UqFBQI)@SxҦ/ |/~ۑF$CP/Cq^XGlR:9iF?F?EEǭW_sdwj.VCAHb1tJw=M<(/pԯJ,Z"5-DŠPo-`-S+i!pPu nAJv\9bL-% F1[뢠0&<^$hߘ#2XA65|-ЇkmҿW?<: bap\5~{{X?K?/If%\ Rw *kFR9O Lׅ_UΧ⥥Ղ$x݇?&i\I6|?I dC;x/cdeFX׼cĐB O^7q9]SXoNap﹘c:fP YyVdU_]!=uJk$0bVWVyF7KZU]R j VI3wv@Z/w=pfo,ZIyo9{wag80F ݆LIj6ݎL# |ǾvptN͏TWm/; yHyf[y0OW }FE5'~m/>83O_hKIǁ[:y!+གCkOV+0LtTV۪rS2op4s7ZWJKs!i +TW LwMk)0rId 14%yרּ(fD %D\^xwq^}u cM&޼#x6:lՒJ0oPRc!-!w-U,L)fE#^KM4˦*'x7(w trĻrHμ[*fwkB^fTx'G[|#^U׎_9!M`\J {}=^bkV"L/vz:hy V 7s=<sBbx ?3@~b 3>P ںxkpvq7y"c)mDpa?a o:ދӁii¿ӥNƿ#8' "8ޝt~u) EJˬ{s)H)dlmZUڷ{ G +cö7=mag<uhѝ:vsA݁@t=9ЎŴ~Z8[Gk'tͰGŤ6V(5; 2?V{]?%izdېG %H BOb/ؑщODh5y|.eE~5ɗn΂*ZFՌ]k.'kj ~\k>mLVp>޴I$qfR/$0d=N / 4wAa`ZzeJ;r&RWdP^fݝ<_l] 6][WsaiKN:x2(b^4ǩq~~U))L 3Ԝ K/f)vL$ xsqA8`TJm! /Ӣ@ @ZsJ/%ECM`l5m&vHGdE#]3RmKHQ)\٦%01a;8C2cH29"RwOǧ;UC'ԇL-:t Y2f~Et"E%5x~T*!+DoIp* yt0ݧe:U+oРTyI =MίZ?&4G1Lp3Q5\ۡ"GO38Q^@pό%:dXoጻ 1wχ77HT'mFz3XS2Ķhb%aaE\&˙]Al ̼ y1MTfy%'IyF a_c Iދ%(:G ):_QEZ=,6666[5VZ.B`1"e2j 0\F"AYE6`bc6W}Xp-+mÀ1 vQ $@DAK)lTmQcCn{ vAf1*V5IoqjbZ8L*3,7LV,[B[#1U|e8p7;;7| ,?)I\%/Aq6=y|K$ "zL}0U;1DdX+?$|r ~ޔVgbfog]77p}pl=E# 3חTs(l[||{n RiӨkNCN>' qf. P0[e(R8- 11?>8Ql"3)_fWw7'T ?g 0;?ˏbaAGH bG&jLX.?s{k\׃ěĥr!/怱xv c\qHhFt H\O{ZPX)x%?χFA{5mP?u*#J6YQj[.e|s{T\H׳٧OQa6jl"3QQE[hpUD(ݣ׵?FJ vnt#l5CGöWIn~K@XhS:~\zlw3 9G8!Ԗ??\=˺Ǥ˧>VbBc0.?nXd+r_՝~1Iר`QuӈIuwkܺK[-7t74PRLՔfDGVNe5ƒGGԔ7D$Vd19B"{1) Ͼ`I?2׶sT9E6r( Mu䠫 A9/9h#A!Ya܈WqsSL%R $4L ;[q-55 f, c0S r1*5A!qşCWT#\$cȤGCpCKϟrL ]6@6یN|+qF;y3zEK]FЖHct_ < '_,>/9Pѻ/01/b!)Kh0O+E9Y"G}">&`!P-{7AG룾Jf~VeאYL/dԻ X8Q8̇[W.2cǚECaVi>{sq<51 fq toUKl 8WRm:؛NIYzBQ ]F1teT ]6&P%Qox:R@Jfi#-"Brjozzsnmw՜.Tj94& _+/o-I0Bԡ/7JU|<BT/ѝ_ϋ F;Sao "3aQ!+"$`ṯO`R{!|Jb"8dØ9)S#EA-C< 粃&<=EùÎY"V ExFX[ , ;i1H)vA{e="NpZp4(<4g+M9/h`RDK#$ctQ-uhņ#ꥵқt>zK!hm H"SƔک(H)SX}7 DEHWWzV#{rF[rd;:?^>Ae/.t=LR*<WX?imDJYLJ">E$IA/> }iy[j-Rvo$HVIY̬bqEu"Ȍ< 5CK~wx_y߅8α5LT\*Q~xE7Ô{+[m'p{_?wiP#0Y&bQZ)-,pg}zމЍr;xbhXDC |Ξ8ˇ~uzZiqM7iíprl_ GOƂ].7_'?@~Ǘ+H/TtJF bot ?Aݪ!DH5oF'+ %VKkktmwDԵZm k,D{\G7uL+i\;f>drm z|xXT"KXdќIwT!Ss!8 '*h /rh5V4=#F;{[Zy_ JqTMѪ{kټ͖l*{j#JsFB(Ee9ϋn OJsU*tdA  wRz0I.x@·ᯍ)ϳaa;|w ߸R6:w&{HqWVf=4sf2Ƙ4$!O/*H#..[ 'ſ߈3ƎR@[1:*.EE0ͦBf d an%^+33Ēh.`Q7~]LP0LòuzM rc4"k(s+^ILM= YyT(vf0ƠdeL /&@D)r1;5&rk=(σagqFZZ3R9BZEg41@#UR,ndd t#[;hOj#Ceu&aq_eQ0*3 8n@c6 z-*{U˃J3US1*p|||& ]\JS0Ge+fgfZP(zNRY,BM$_<B)%J3O\R)9mi =̡<>5 i%XE)3-ݝ}rJ {u^ %( ;z qJi,Wu9\cUF?5EDǹ0"w_[gX癐A Vt("۔m>F\(`eIPx#XuF^<ߝ4-qWeA>"ȉT39m 4%31!31?:#ńްc[ ¸_CmkS&1DVǛ,w6n W!O ~+P}#ߣܖ?E\ύ_W e͢uB|Xb>,k6cGɉ'8CM>mIC+p=~Nݜf7MNS6 G^N#o/Oi _ki/}b)v.KzSlxĞ{#C[4Glsݓ/X sn}n_}!8pZlaj="Tl◔hQ Dʿ[WOtR;i;lalJΎߎ,lџwԍ!([\ ̙AwL}>G)6ÄbL#|>$zn9۾|dG%8{Y'˷ϴ )XG8{Z-*@JBQ9UÐ */ș:ZxW3gB%WߖBWÖ(lR3qz/SQZHǢ! 4?Vٶx(]/X$?4/Ւ6?*vr%mHt("6NޢMSIOoQELV Lr[;y'3]t dϵNN! YDFViy{iC^Ȼ=I)K9ϐR˱}Z02A[ K-!$߀ُ$gl;>-(in^3-,*ݺmUx||٭(FWas߻;*ö`@ՠ"V4zr߱-bMl?? T:m ,нA]2ݻ\o}fsIxi>{j}nUA/4BS8ՍC릑u e侣u;{)UY|«DVb!5zDq`nкb2trǺI h-l?Ѻ!o|fF󩮧k0creAR؟^+x+lԕiƇcJ OX ϝZ8 ELU:hV1H\ͪZhMcU;6}FwJӭ &(,6"MWI 1E5>֝| tWdc!"GCNS̈}.űY4yĉ K|Ri2~Gi#I[Zw>{ -kQ/%+L/ sqNk'r:DpAZ5ګͷy{{9"'芤" )9ǖ))) #HXl)r`NnVp E*IO?:W3O QP%G W  qOG{y|wwk߭_.b.ww_gt^" w>$ΧP#R.'1K}͙2'R 1bnW&%gIՠlAkY:ȫ4o/ x $ԹJi5eKrua~ vQ;l!y}8_<j 1D~ʙi*՝/]ӰQH~{ ϘBڿ>C[c)h GfH, 8f~Krx.Ҍ$cd tPD侣,6T8D#ǔփ9\JL@?i87<!n9Elj"Jt; Jǃj:[M׷Ҷ>ַNմԭtJpF 2a5>}Ўuhh"¦zJ;x;aljXyzw*gCb"?AcBPLhNUAڝx cQPfmԼ*/gޗvI}R׃s[rxh˞G.Yj 2>Fݝv@ MWM+MHRކ+??ٻ[Vn!p 5HQ]o41!`!Z.&1!! 2{gZ 1J+jVYeUYj}>`Jۭi椧9]<}Uwzg@B8[M+cNMU+zĹq ^r:"&'6|]є2ɢ'ߠ]n]xa5 Eo*~3n|&η|[o*i4kf޵6r#E^|?OY`w |ʲWMߏlVnvZc!ر[dWUETcGVRE#d9Ơ`ƩNH'L)cC ѧzR&Ar'0Z&i@6Rxodn3el'ߐ(:ֻİ4THUP db 8Ʋ`w8X8`S*u `Vu=Xe7.WrT F {ܕJe-rs5r\_#>_= Ly0RlXRG@LE%Kv df#ۢKs_F"=o;H(`t9bv_*F Fg@&vUե- 96ځVdʀ#FW q>/T< `i gϤ״m3t83CNcu]ZW`u`Y:y"E_nkAŴ7G]rb"/Ѯ,: EX=>rTrTu te/ P6'XG/6w7uuQC Vٰ<-12 QYI &zoB9H ڋ)=ȩ>)k I ^i b ֿ^cb֎眤岢0`= |٩ecQ9ܿXM;Fw3e(L3~8A'кBB (V[%@ MX6ps#1 gfZιE+1=e\;I0T`ûQ1b)ߕ+AAs"mg#R =5stB;S0gdM`!g3!DjiVž2lZph,^ˑyː# Ph K*ԠYIe`WTQ^;@zDμeNא@ X Zqנ9j.<Ogַn, DRMQ8S"lקO{Evx60/|OOv٭[D]afnsڻETƆLiǦ4qu[[(öHQpAa,Ծ7,fqM($%bͮՔNwHFeƚhܒI;@ @lEZjjeOld7P |ْT-{ ߳l [TyMm&AtLmfk,;LT *[eBuFRCţgBX- ؜`mڝ<-6KӯnRDSJlI[-ukA+ h{2ҬUO%cA|,NIF;Hl'B !={v+Zإ"6cr4:1E݉f2O1I cАSҝ.^u+imE ٛj۽WO,;,;'=jCV|z-nIdv+^dN@@ Ldvj{EvD""+ ?[˭^^Fw6N|I!ȻP^^~w&;PiAؠ޹̓o֓[$F *C.v/ w>;wR7#%ekvDtzQOzl)88Sm;Ņl\zNI:vonг1C!·˩[=s&[]8_rؙډ( _Ą'b¯䜰3!ZcE"9R @ $R X8ZPʽdv27aExqo0NܘjY5ٱyͩlo|Id5H&oYQ68;92[O4!T0ϷB97<xRS0PN|2Gk+9޻Eo?Fw:en .\8Y9!D9fi9l$H}ҷYĠzrOqbZl*fV<|Z\ 8*:rr⵵dRå^!J:Ξ1V%8<,lu'"0s,8xCy1уŊ5X+'a ,d7Ξ7iw`br蕔vkIs*% WT cpֆ1;12cz+H(<(G@3*TrGTwղ ,{ێ Hˊ3{/D-Vq {>#=X $&& @Ǘet& 8S׿K}CE6 JW8֫E8$+p#٧ixŝgH_\$W*u:h`mzv!n+a~?[ZC_> tڦ4փ3?=@MjE?@B kb\-zQKܶ^]@K*nwu`;g d,8὜D,ц1P0+T$Cn|.xr%"QǬfգz(pz^]Nte2]bjlcWzKG m-`zD Z;z ݿ_mU Wpt O Lfag7j^:ipr 攟Q?B*q -RN:UvCZrs7Ab2G$Y!ފuZuo(FeoNu7H^ʻss8W߰%$6ET;PhS󶖸T̠`8;[t2d{QRy!WR޻9ꘆSLVxumjڐw.Djke_mdqhv.IMJH%K|s ak ]?[`!xw_:B\3=}AwE0{`{K㻚}Y)?^:jM=s|"`/'qfP9BGjwq= T.$c}'鏯<0/f?u|XPQOKؿ{LC?^¢~fOZM|$pmQjӣXkA@W~o&Apis9jW_ZaCRFӫki솝E^]Ik-^˜4'@^*P, V n9ȤbeU5E-e\.*7 ӌwIdbA;Xg8`!7[*Co)g|*p(XbSLi &Okc2C4>‚gēھ+H#[/ݳT! 8#iߦ5p)L wX$2(-Y>BD%+=64(jP8`Zr༊-sGEpXԳAy+`6h9Amxr]ȇ]jVs1{x?nRv9Xʇ+bio~]<8R@*1n 7t 3pX i˨:hPof"xP{0䒒i61eC]岡_é4dNG)@sB׿ε׉m-)l]i9+B}Qhk 좍vYC`f2j-I1KTuD&\evdE0xE^תkε~cҁ)or+knJb|n{ŝg;BK@~go[O ʩRVS Фptp*4so?x`,WDKf b4BgD:#:cYn{1a(xeBCptw RS k3e{N;"$QJL$0!, `XR+BD{1 pLj[Ŕ $)aqJq{1P`xl"1 PX ٠{Ž|clL6Zk˂\ JHIBذ]0A0TE `F#f׹c'Ĕ8х[HHzm9y#m#d^3թS)y9ϒy),䕛hM9U:y7AGĶ[\g--rݱ)iSlb+Hw:v, u_ܳ|B&9Jv$ѹ.AH**ihT(yƐNّNS+g!N1J'I85JA m$b,ۏ&lYmWg~kUug/ 4[v#,HCLrO޹ ˧;Wa1R2hҘ۟b[=ڹ"dpSxicmA{ˢC] ƔTwNUmLULHն*@1ZLxr.ْ3`GLc|t`?u)w;!2a$ʑ2l!B1i];ս.CEwPtEnQta,܀, $aN{3ڹ-j, :P#Z(JǼ=JU8GUIHueJ2 RHKrDVK xf åf%J|L0=__Y*CJ$ E x2++Rr* .2!,hrg$Jgi̮VwUѤ+1ȓ@ƲplSlJi.aؼnU_ffҶ-$!"Ђ"S)sjHrJц=kc0BF[d .F+\U/9lxW^`$Ҳ=oEu疱OiЀ.=e e6>-ۼQ'DV?Ϳg?p.p:Zj/_nVM^6UfZwMSz8&K9ӝ+.C4j ܵۏwu!1 @ $>8;FksMMb2gUHh]w Nթ!?zO3j LAw@3:1q.zM]/H_ Pa2{m.5Yzݪ6u\:9Ɲz6:RwD! SkUxUx@U򙻧 {sIе򈙦qwIaaa0OǭO>hϴB0oRZŃƇxj9LiCkph7KP;R}2EbEFCQoVtP}NPB^FٔOnAbc:mxYSZmwKa!Dcl9Rgr <6' Btv{/9愘vu%"W-0M?t& 1OZ3.v6[TMN?jU}QpLάCߎ3=||fl9ٙړ=ٙړ{|O:v4xPٝFYY=f?1o: ql7X?YֳvfudF6-> NJ};U_{W6krDsژ/gtg6N*BUa&1T1U8'hUQwa^IjXޙԺ}O_'-?OAVx~3~LG3lS3t$"ʱX[Lͬy"heEH9;{}uic"þq>Ǹ=!;fx!_4LCs[< fwnRC?'NxA_pV`CPT t4IxK@9 Ȍfͯ[[he[оÈ֖k~I'E%<߶[@KrX#`'`P&bC69NU*}Gl& \!34-n 1ATy߱ Nx&. B&?N22vd b7+f +0/:Ŗ֖֖֖-i /4-$2B%Vʊ\+ sVb8ͩAKD)` +jtm/,B ]V1kZOonm} OE ('g.]Ҭfiex=>?T_zpdc؞g|yw]6=sjqI lpǁ \?:K11c؃CӒ+ , ii9mV Ve*GZOK!J$3%B1h&0 #D!V EgU0Vrn5Y/6!2?܀3'@Gb;UlwV{(Y] eg][~<4vjF¸|EU݅-WG\Uܩo+zw| z0c3Hew%U5A 'xB [N0FU &"km]0?B0Ѹ}A #vVh1Vf G(ø8/rf"YAaVVʐr(1'ReRhBBX .h%HMjȘ!St]RKj׻RC''_,>6mp8-)~_Ja!DclJ4e'y74>X |L'< ƝyB-rmlS&쁁ԁ(DH|Ĺ_{|tW~{pmƜ0O`xε~zlQyB\F5D@t佽WyU1; )[R%#YƀGVn 3-Bktoy zOw.qQI?uT`2tR;%/-V]ꮝ*P1$jX|ʬc DkĞcVN=ti${x5\#8K&%pD2)@u+YU.Dс@307cjL-@LMRCbPt9 -f'GH `-g<Ơ"2Ơv9|iR@ 3 il-1-9&c3ٸrŰON`>9ZlL^B,yI^ > C=u5I#{}?RBM1¤%r7t|`hMc@&37ﲃ=cr 7' C~sJ.YDGmEw t( =Qfys1~%a9&:69 ɘےI.z'|WS;v\M%$eX f9A\PcHRgP*A⼤7 dlCQ$\:Tb'uRX_c'᧎rH):xt g{|?#iApg96͍ 6 x@V}`6@l$ij|zWtoRTI<]yS:BxS]$m!V$.uNO d QC_"VbBeėNׄØYԁ$ԬeD Q8%4&&ݪXLS̤=;H8RKT(%c KQ X(-$ 7rs93jcxB(", VB+(R$@X R"fsr&LPň4Id4c T3tA1' 0(̸)):(GFbq# AHwlƶ‰`u*XgkONMMA kPN5Ղ IV[-h@@ q܋ RrvvAPCہXi%J-5ȕF`g9 f ̐ Fںk°沑H$TPswtwFXb΅ܻs`%2Xކ9 L^y-f8wFzڍGJH3jr1wa_(- ]i)GL4ӽYPR!Z~% 'T85\7_kqf&y &LVB8HZLxz:1إOc;a9{E oj4mۦpsCLٻmlW|Մ/ifXl-ݻ_z1(*n;-/);l%ܨ(D<<90C|RX 1wX H!hUXix1a'5V q(q(v}.fR̔ 2)DB%`3%Hfz.((z'|DzV4x斍L6\_ O mز9+ _KV}hvUv{{Nͮ63yF./ j1$3 ˤI+4B~u~ήQTHBv4KITB@%VJc *M!9+JQ`\0:[\2@WvBe` |e/si@ 2 iƘ1FЮ()RS )JXb۫܎z,,P^Ra =3HqJꅖ)i:so%Tdk;S} |띟Z혎[bˁHku{黻ύ?}/=dԠc"/5lv }`^| ד^ч畗U0A:|z=[ u7FluD5 ־g~~!o0jN)l"\T*F4sOC>E-뻗5sVx,Ψ'g@6SXZc¢MHC9͕-9oxx%l:€ۤ ExoJ)Q@0CPQRxpB!90]7oIᢂHֳlO,]V47t𽾻{B1\Xu=xceɃՕZ~>N2b04tp)("vڱXj\ ,v=o5mt^©M 3C^Dy JF9bpIs"( HKa@K JP;WRC"}bliq,[Di4\)@(@(ϙƹBܔT@u@KR2WVKZ(O_#6[v@m5A2-gt<ވd9jr3\e y.S~ݤcB1qgtݎ ~Q-qSH7`3Ą("Iq0Z#hu AVp4KZNX$,z- VX4 1_^,{Z473BXط~T춝ѥS%b)$m_,L"ȕ)bu ݬ 3b[FAn|TnZ9Q]^ẽL悧0Y>)`sR#V9Yh?IXw2bAZ;#~KhNę08Ҧɵg7E,HM}:`9z/olMq ]/.XCQՠH$t>ۍv|@DTwD"z` `/!S_y@S.{hlx u I@^7?;ܼܳ켧7Iaٶdot.,֐DXOk[=-6GF|ZeVE}ZRm&Si?I ^0)PP]}bIa0 w+$2ǚzo~c\}sI{ m5g65/oC^A")i%Ϲ`:OWTX/Đ:ajI9 vtJ):uBcpv "j5ե9847ZS)ؼ93M6W7"Jj+`XCĥ+`egMJ ı^x0Z)T ?xO@L'(̄$&T pBٞg<<׎Q<24l [bOIӢ#m훟>5EJjjD_zjv4=I|YY2/9C<p$Q$&I I<vD)/o>嫋y2G/\vg^ K|sv'񒋳P)JX(,J)rV꼰 J)xaJ¹ʱfc42 BI*=U _VKfVo j=r_tOQqK ,g2ZX.#cC &Dvr-U>mө_3q?I ucZ8Җ-K  hE.*Xܔ Dђjl܆KHd5몺y|]>~Cr )KDunZ~?ǰ1??a %}3e-Y,vߔlƝT_ÔC2)؉{2m-Z")ؕeNLcBriq)I"4ŧxK$c{v_OqP&;KV-Z*$9 -Υz*[ǟj8}amCLtf̓[7N'4^`0␥pfk;o˫LZ~2\2)2{\T!!16oX֒ {}wUYH\?|\,{AeVcRͯݑ pypܱ.g[~Ya{.fwKwZ)xh~|yT.͋;ˤ+|TE`Mb^,Tf_J^|_"r~{>ǠٟWG |gJ.$+7!>8S Ī;;q eǹCe=j7E}]OIN(M-h9\c82n;4A1`MZX ȨS\fIDƄḰ]xs9K{J#ҳ}7b蜝Ad~j ^4Xg#ѫj>F0!eNҪuoCqUϑ=8Qz8[C޵^4 A=?c})M;z'ƃ@ApUS##ҩsn7B֝sO 921Qwk?^sQ0}m>?Wvop^Rg,w,Y&EN2[^/Ο OBSBo}0˰E1yiE^?%%)FZт0H , U69˽GWUZՌnBt1[_@s]aTvm~xo7 {Z!HQ_|*F* c,+Z\d$\H4Mθ F%" Bԇ_:A@1'crp2M$Ct;KEzCI2ip:OhC-T SɡQ"b~6vF' 1,4Nw 9ޟP;HH 5~\s!,Tsjir&@N{,e2VRV En_$(*%4O%*tDa{(A47_ͪ^9&6ٺYPc6K, R?\[OUQ6pwVgj=?|.qer矟.XTxot?q~csusyfG, f˵|x}m3SUεZ,^G\szg`m _yvs'H(xlj.1v3uƦ2J!'bw\^LXCFR̄ ƎE.).R H^Sک*7;!\ HNM"@Qɾ '_8VJ]͞:Ib-oS^Ǫٕ>f2RNȄ] dvVSe1*PIdm=@[k9\RNh8BD\X.((82\`6BPݾ^HSZE!9+K] -|a8eK9`עm[.(ݧ]@qƯ/Q$G̮)̆"XaɌ7j]-~]S>(1)ʽ9(ssJTp>tSLҿ9!"$ckW/^E8$Ũ3@`)Ewŕâ3]-:ERo;k;NKJN"vL̤f&">n3 xa P(/z/1xħ1v:AMG }c4jX9p=1F f= 5/>f^dYo z mz F Y51|pSU;R>bԊzZdYzmqaKsncWn8ux"]_o&ȘN+U==sTow (蛣,<|~p= -$\Oz( ε391a[y0!Ǩ!wmehg%Gٷzpm='+NV1;binYt4]UD(I^9fp?K.^/H7Δ^9Lp{fn tE&_IO>V+ >[W'SA3:˅.4KpaU:\ZTh cH½3j5V%BWkĺڣ,#]/$k5^0BqT qqt' +R@@'Fhb.t83jh_9( T%T"$Pc^v {|x62e{*MO%|z"_~X-o&5oNEY—c}{U]Vow8^8lzy774l}N~N7ͺ9߼/S~:FEr#gլ7Q5 =Rve6;;`6 鏇,rgOg*)+&_MUu$ $x39["A8Fv;8jԉk$ цLEbtU ]n%! Kw E!B'F22uY(APMשky !v3 q" Bn(7[ +k+B(ONYŴPJ+6&V!&EMB.^o.- 1jnwco>81UlK~Z]gvK^ r^nn8y9%a^HI/D3cyn v,rLñ3^fTA'r=ɉ32D$S.prR׏Fl=tPԩQ38Aܹ!/7yǁ<'!qv-Rp{xS|:ТoNhfiƁ7os2|Nډ3~I|1"FƧ";q";Q3/y> 8yqxvx6 vq3^Fcىa@1$U6UjK;1] ԿW#S~+Ĥxb{F+EL{Wn :W` =[^@lZzCP(0>S`~ cy?NdžŔȈ8-{.- G(<SH?&JhGVvRΗnbsHfwr{dsoҎ%|ҷwW7jᬖ@ xQs-d'$1cmRvbnMF#o~`7^Ϯa>>??$~#Zt$+j's"U% 9xL7Ot;Ou1%HVA#FLΩZ)|Sb@[uVb&~|kG !Vr)YeO0jnΰb=6[?fٽ{M&ߐB;J?8^H^0?$ fuq9ð}{;ZA ^YfR&pzЃA?լW˹o*sB~uUT-Kk6Hf'FāfONKEQ/ax) qh6KL[zX'3P2+XB彰>f7g'"35 ߻nn]k29VUlyz)ϟA~/h`!E3 Qԙ)V `iaKRcJ!. s\}u4=k>ؽ (:ҾDxHۿ fAPH0qӿ` /QB$j 4ZūB$ʭm\dIDƄظq kȡJnFFu bca5 H <8 Pb BdTZN/Rͨ59RY0!: ;rn ۙ޷ɥʌ49pk".%R{"bkqk >Z6E!"kS/AFB%L #ҏ>nn\tJt(NJg^3h=Ehc6\gl3Sc{s꒸ S҂dVqwje*H)FX'/!j~7K<_^,|l:f_k bё1^6Jq_q<:4˞UXQ@^S8W ^U+Fc^Q(ɞQ^P 7QL?7IN;iJ]*ezrv!I1DdmͿwOڑ.;p p?}qyCZo\$1]gɶSS$-iDm\nE#]^/H Y0+e3A)Xs#0#c( UΤ ) UM()T5ukN8q5fSzh2Q)ikմ SPp2P+"݋FS9cp$̢v9xDSa!{! 5߇IdG)=RΠPU: <(WD&ʕ7Q&J]:.cB;e 25` ȍ0T,֯%[,I); " Nwͮ;;Je#fV)a!Fد4w3"*RR0 ІQbEβGT35іb*R~c!]Rm+FAN2I&-I4o0DzCD@?}FZsv99DɄCPka=R$T9ڧQt樀T6 \Z٢Jbeo|8F.,ΆSR(bܗ lg#Y8 >}t!b5x(u{CppNӠŸ&i vX IlȺMb#r[#a.xnJPAa)l )r0$p\s? tӶs)rO!Xzm>BB' {B*]teصW-Y{}N.i ~zm0e4cO\GTNu@KKyAKLǺ(p8#%>TERB&>;saVQ+}qޣXnq2[_xZ7 ٥L| /y:i&B/>oVrϿA;l+B0 $M)BLjdd0EF ř">C(-Qo0 &&/%28gG/9-72h4ׅa"}$ Viĕhr%#0d@`ѧ5CV9%<ЀJG4VCQĤۮ$V: s+0$O2]S_O@HqVbfX0з} po|Jܑ8m(.'o>~v`d[_BN) n;ٶGoeAH,:J=4 [;+^32FM.b'nY˝VM2Wzެ>mj-\^&,+'ܙb:<ċy4$Y9oO{#品#vH\ٖeOP}{URv2Nصv?T;>ɖ9/4߼3%P~:Sr#g7ѧAU$9WnǫbВJzk)HBpM)BGi7j΂3jT BD'uv>S[zHvBBpMX-}kϨR11b9O[zՇ@օ|"Z;PKqhy7 -G$IB/lQyx5虗yI,RbiD*#㋌Ȍ-NR2\G8`40+qn? PNq od- P;I"Ò'Jz? B\>x1,7ͧ|rw{#`zLf؛IoRhG͟K-Ʊc)Ť >l ]9iRF7 sreAM0,0u X\0dQ2%2l'FoWf|S@Ef?}Ys'IYeA7w@'c>hyuyUMR9WA͹|bJ~S[geׁ)1.Mhwj񞡗}Z~0J+p<%+/hJu×,6YD}_8SXPХx0_RL 'F=[xGm8OtC! )f ?sPUFqU6м E+͝E\9|^_a-}ͻlq:$b J3R { drvI+#uZ;*j!vY3au5]P bsS ]BRsr>cv9բ^=l׬/y (.]7ˊz[WvI|: Q- |6*y f@d c. K; Xf~*j ۑQ3=ԓŒfS7VLxMyj:mCf o$&hCCz8[4QzG~M;C"J_,F; ~MG-~˓ 1Oy\mP+T*]Dlw?R ё{5HW&tm_(%pfPeJtc(+zݣG1:z׽0cx:HiF^y'Z9H<zs]a'OcrbM+ h8Fr 4WȨ 8ӿ.8DP޳dV$"LtS}8]TAh+5X+R"*,gĨb̪9PZ:b %Ax %s±rV ! *%RZ#5j}-pk(Vo9EʩpF8UNI+I/J}B q '!>'TR kIv*AHS dPJRJ{)KiHle(O)OuS7gYni3kWo1QfAq&;q_BneՙRѓe:3[mX8D즾%Uz;¹@k 4V ٢˧FdM#r]QhDG& t; ,9$tnsk$ CRw+yR3^n[=`OJ[`:*.L\,GI+>N:jcٵ6IQ'Q>x>xY?&ζ<#LtX41>V(1ڦ]DD\D \0羴"l1#F9s~=zP_nΤ zKB.#Zi]47m/=cd=YD!1C X*j辣\v挺7J!Rc{+?Ȃ\:OY>?wղκwehD.$D>ԋuJhnG"q|x81v끩hYG75yr9'9!j< Cë#]У`'($̞l)ek?C?r[#Xd= 2d=qz);a9gÁ O6!?)G1CI?Tx7 y-k%r#b1KȒ@"j쉃/$"[O{Hy_Br}{mV7.{mZ3Ķ&1Ukݎu[%)TPx&ҽғrdžFč 3˵Hť3ycGFD$wz0c^#SS3!3%̆l} $"'Dr6 X'D.\;!FHHx_ SMI%%U|'Ր>a J%͇LkE j@8!BZ+Y)RT8wXA9cVdZ/bj*Z11Ъ1 ϡછJ'6!gc)' .!)^[U;vXu_2v_N![ҧ]KϓyU*@acۄyP KtW}V BUye"R\5Um48%]uȫ6D1;vsvʜSʨT**֢vaύC1g&N! {,Y!L8 kxFLF_]>()XI>̇WhZR\u6)U ( <$;}+#|{^$Hrjݠdd*NXĖ2``&~ 3o?oz7+)sS0]Ɂz/L!]0'qb5p6_Su2L>B$'HV:ODž6 G}N9ue~qNЃa7ǚ5m<7C7[riޓL׊ ]?LC>=RU=qlR6lO.J㣱GZ);vYͦ;XYsĪfFʚ롟4㗋kAK =[wEk'Gq?-(ﰺrMW$VR*ܺIzJ1AuPQBpEOi 9iDu9h}ۭ>hL5+׵n}+؜j]';RAn$ɕZLhLBE؆ B5WeQ|/j%ԋeRMJ>N2 9+`56,neMj~i̓_noK ]c] .܏ìbtj´r.oo__y}^q7| ˻$;62%;ڍ nmy#:mǨg{%L+nC2[h+_Źۭ-rDu9*;vk䞩ݺ\Dd*P'}; s"ӛa0Izx:T~\WҵQ͠}0ޮM' Y'󷫱O^#* ')4CWC44J}%I|,?*FCYz#NHdf!.ZZLCoOŠ|>DwJjuGMf؛IoRhG͟K-Ʊc qO]6zWNq)#esЬE SY#A:QK#\"vyEOaΗZ94[Xd䴏E̖!ڂnp-c U"B4ob^hb h=ڜj04Eu@ā(h;wP%xN ZHS4ұ "PHwL# /XS3#DpTx-  Ls)\4<_հU R 3N9vQc-|#2jŴ%m`0w6R&9aAJ"AL-~PDjFn(wd8LqI.f`F0!B$±!‚fbAJIiɵj͇QC5Y@ѡ"EFH'_ ipTP 8[|p$r!Hem`rB(ck; Iwm"0jT03,j J`%WKx*@ FؖTg4F-_PpTV3NAJ:b}5 0`R5`d"Ms9)? \HF7c-GI VNȂbDت[jnH5'GrT Ļ HOבQǥMuA~RќjAXMv- 7,hXa*` xTy<`, WU`K "rlkjA'S0H <`.E7$S)nNvIO5j.AsO; x4FȒmp^P0G%} ^KWjA4gS͎hUPiMvH@* ϝEPaT+"nx&gYsIhcX|%(͹,DZ3q&=hv1dκNkU 䏧_P i]2sCi1 HtB{f@djc-޳W}咡|!&ȗ=j׊7v4Hgzm%{U] \KۚM_=Mm|zߜk3:Yk_ z ߬rog괓ĐIn//*xNǷ6F2y#?/.uy:M:6h]<:M^xm󊧨i\Q]VA8zO3^x4G>q7 Rd@!0#T]%{O|dl$# 2PcZR C@UJ$!F ֢),WH=,L(O;̽I2Ei=\MM9D3|ǰvbS;@gT8|xOe7S^R/pP0y8y@NzaСMyUCXm `Պkc+yK)+$EJbh; 8L4b76I%{*TMjh"i ]G$A]+pe Bs 2E u@h0v0#:hh!/:s|pʠtQjg`֨vD16F@ˆ3/ԡZ5⻕ZHWF#UHC53R13!sPAl WxVy#LEy<`QQBLո\Et? εopч凿|w:C˕>GǢ_PMc~01%T66]$wR/V;5x`s}t 1 hO+Ezé@=X4Z)] \+#}=cBl.'ɌsgxI]ٚP.ްE=.ZBq5DQ^k&W[ݴ8H-l`Ph $VF} 4b<3HQD}˪6[S]B+VZсmKOe0hhy˓ĝxu(BCDJ]Pc5@>ptJ4Zƭ@ *!0!AZ''q9&}+T$Z #)69m2j_[BVAV-Z)l/Tx[kHS]Rާ{9& )"쒷p 53@gHf |:cxlngi4E'm9Mûi_X O|R>D@lʰ'Wqe齋v0F Q $2lJOq/ ^mvz3}?tBcX"c2ѐ79w:?;|pp`W9;-nl0݃w+47_B Ֆ/NƳmbyP|G rƯ6,`>|"Z%SFN[- JD}])3toڭ>P!!߸ȔٛgWn4(wntڒ[}CvCBq,SpDrxAxݰ[-GnUvX36Â׬( .?7326OSNm4񩁔"s1LƳH5wZBykEHd0o+'v@qjx./ Sf=ʞ7W?7!E|o-m۠, &) &pӱ6ȟ^Nkn3JVd"=ԗ@ǟiyr@wY\L"Oc wgLWT۱ ޫD\oj,I|zaUCq~U [^.[1J~X?N;iqt~i}xӨ Qcf'7;_w>)1y|zɹǾw^>̾pV|<>}$=YyȤęs~s'PI.BI)8;PL s8iF;z< "4KLQc&~R] rWqY--K0j"/뱕~ɨ+Ԝv;,~;jgY,,lPA(laIG5zܗ0E9>C(6*[zH:pXi!?eAF؈};ʌ8*pQuTQm{_ZL6C2G@nraC-/]{L_6Sw됫ӏL{-]K6.sD7Y؆!`GN^@ncDIlċE$81FR9JO.((%1~`0$ ͳi,荡"6 xmv(;06Hw}o-8%?.և/ZmqZ+-T3^6D,bX/;lUO6ZHNG 9Nc/2('A\T\T3/v-5b pz`k0 $%]2FA48 (vҦ.>_Znړ߯r ly"uEkuC`xtF8=լfp!ÑPa{Џ_-ݢ{3V R߬4Ed@kb16sZ8]z I4p7Mm8K1I{}S%:}j!jOw/<#É0ߗfo1gin$E%.|IiOw}řKWmuwl&Gܴ4)~c[ I?߲$$$ =?!r@|Dh3+Zs.)VR AC.H=A_^ vlk_hڐ.V-?ͶmM`J)8G3 ^+Y%rl_CB:NaMZ56Ӛq-da˓tCx#G$Z+6ʃQj֗J)JEܔ#^HD>cM'ta+rT r,+f$GAcD,n]|zVL5j=\ںh (zqT.[^7K=Mgt˥H(Ʈ NwBTČ#r|P.|xZ I{n<`Yd NF!$Xe9Q$6:1f\ $xyiN# zFq{wDqǎBČѧṁi1le9%67W!/BQTj%̅ZM^xWgwk j3d6u*[PVQUH=?&Հf03@H?}P 2n^2pixO!@)1FpϯE TN@3OI0b.k&uM)5(ݹg"RH:=n󗬰dyd 3V7)Ƃ1ȗ0ye 7e6\1 # #Wk"țS%a I)0v`1KVy#yt9n(/]]H$uǑ/Ya`c_rĹDOjh4JV+@F9Jγ T.!ggoP o58 OG6&iUhA>,m,mh41#'ўv(D~O:!{{~D֔ܛMvo֏fGaZЭE+$qc%7K,>  ru ,nGD E Wm>v2;ǷZΞ1*\ *@kBj븡@4Xd\h*y| 78v5{xYemݪH˰ei[ķ*h_.=לn?-MZ8F/JV!kzy,X&Q9??^]OǛCiͷvS<w)T/S:d~+w(f gx'O <>BylMc]R/;R:dP(TJSC # k"EH5Kku.wUc\Z)wq%( @ \aPW{F/ T؅iPϨ2^Rʭf)ȸ>y*b%[$QTSÍ$j A: jdT{X:Qm8rkKJ4ֆ][o7+_Vx)U)> f0xk[ȒSX=FbOgF؀eiZ.,ɺN+ P5@q@F`LY1W29(EPl͒BћG `5o6M^MRf[4AgZG#!P/0v[0*%* ٠)0ȜgEK-A,˦\*3$RQB@kтE1I:KS!HydĦt?4זe&v{y+<+)?:"RgRj-9h3A 5;NBMlKҸ+&4;E&Z3?C`KSHŔ؅a#Dpv [Ջ8~t QG{2e $I4"z5o#^Sv$P71w_\c@#dr?LfY΢@rӗ)-; #.kz_~edgIjp̈F &C!AwЈ3ݪ)yAY(6 rӟY6u3Nr=i<۩yХ[ɾ.o廀:lJ[-3=t]#V>\R=,&$; ءKR>qZ1BIäMl Kgi 5# ^:c -Z&~z&bkk[K> UL ]Bi-&=٧\kɇ${`yml3c0K!3W0xGd9Ww6R,pd^K5{WMK3*I%J8۵WEHֹaKUm=ux:wY}tW9@ij%ux'Trw%WP˧W%\UIkKB-a4) 5J X7D2ExvDjwVN<$9=a6Jٹ,k麵#m 6(۩4+s&30\ٸ$}qkc LrtO%Ak"4u(H* 2 lAZygaH5[^eu]~58W{IK"_䋻XX_?fVD)G.I?i jPa;A*ݙs]lGd{'_:H ڬN&H;żoOR i7,[u/Q9 y^]})!;]&awԺ/ },>IxYy]ZV첆SFr :.Yt*ez6u{!#F1&bDEv{ ׼Ѯ (Yjx*nxǟ\ÊyRyv< w5uo-b;hW]+ndJ r=Z} U4z !2-ndx$YO½4; }9K_W7>e꩷'aۙ?l˓V݃Y.-3[vS~_3kz+ᩂ=:m"Er1ˌ+ഋRZӃʹ=?\ZFUs%O"39uR4x &dlB߳Ct.qa>_/gٗ^2Us\l}nU.y e7֏/:SRߩw XmT%ϓ?E)rqe;_9?}sҟ]ܮE~gatFo\jAFzsronJyb<{txsJ|yc ZCh GdXER1W}tP+TI3EjAO~5(sQJ Rfe}HZhuy{XoZZWzkie<)s۲Kԃ'Ȥ$!($7>0D#ͽ b赟ڀ-Vgň=>@$(ySIEGS.\z!Jc&:Qmq_z}pQ#0S[ЁHOo%+{mMZemԼl-? r"لJ24w9gɽ5)cIjdz-QJp?TS6BdSPo x2+Eda[g%@"#d 3&6Jh5KUh >kYGXܲ|6%hgoރ dB9=[5ky5aJ:ʴ%+6&3>xAC&6j_,67X%£6 L[Y7De!fc>>$)=O5i}iVnTuQW,E7^;s2v;x 4GAWz/+=aTn vp굋{u;<%J4 3ۨO:Ι)w0v'U$}|Z}jFxT}୨:+ϝ2iN4ț\g7 ukuW?~ZH7%*'•MxI' NIG|s3XHM? Bu#vb4I(d(db<Gl%~$b`,O510ݮuh%HĒAy><ꍔ@:pR1:?LQ;]bJ>,(-PwwG@4N!DP)jіp~9)TF:¤!{T>iry6aDk۸Q},IІ`S`!@ 6paG.-1JX b Z $+W& NZjuKAj|d8ě1_cG5fk3¹gSzc$lW"1iT0<T1¦]I< h5=|Z?.lq6˴Epd"C^WnK\O .qWݖjl{[Oo23U:Ӹ[pm%&e}(,ڇ\:Ynױ4a" (:)ꐗzR@tV zKOF`c)m%`[B;4wP /8 u)kaIhkUlrej;KYOokFشͦJ!f_s"k'FdB^6$V"WܷH\s*(2]tl}qo=yv $+#ƀaB>jh_WGJ 6~D{aA|ܷ cA=Y :ՅS` >mJ%8?mjF`$ޝ:79n"'x=ѱaTY?+h[Ӆ9Vz*:UDN4>mΎP-`SX-Lm+$_pv' 5f1X]T=\{ fD Pug껝Nl-(<$9tw8閪meNݰq#c[wf倓k3jwmr {M!Nkz$Bl??Ka>^xC*/ -;o<\v8JzFbj̓$xɘhhѲSeԤ1Dݲ0 J P`´ޮicJҪu$v0>Bt,lK :mѪv/v'9(BN1bt:IRTw=5pDkl@쀫*QoxǟZrB{W%G,_$ue.X>g܍D;Flj8BlZm>|t& rƘǙX/. ,CiN{DJA b :>ģ="P zp.k σz=}Wd1}ӛ󳘿~ۅ|s{O:~_t}Eu{5B>6J[odBXIE-P>H>)LR (^]q!ޞ,S˴vt2EoOʍ盓R݂%W#/J_w~̃.Ro( .bZBgS9{?BW/rіŊW[RdLGbF^6ڦc#BJ'U@dm*1_l.;ۏe'z>/,h|aŲPYSI;j+˱8T)ͦ\>(\mi"^-}`]VHU[n3K>2xPy,lHIwBjǴݗ+K3h{Hjp7?%jY+ˑ,W.6_xUb765h`=?;]1V= "s{ͼlwzxGm] \y|zUҾͼ\hy9FT7o̘ѷ;Ν7B=$nW;ɵͫ#'lySwa\F8(wbaKZvW߾V4Y]UNHA1zcv/4/}-ߝia<!2((cAe bսچhP/-VA5JR]dx`үeM:`XD*efVx7\FuK7ؑvU{m05=2iXEpƈPglԧP[2KkRJ ky=11"z5U*2ɣ }D8&)Ԥ11qkcS̄?;Go7( ޱHU&2R_n|fZi &.DpE2Gl2؊ewх-#ZD 3A]h{#g^]u%@N0ˠt6|e=zD0)M(4HF尶(I!b)snZYI+GQbM@<G/x^v}{crr1b$C@}oA25q0t0iρ%W =9Wן]koB'?+zH4%YJAE! j-6d=Lt4Z>25tl]`J؉Ƕ۸*l[sʹى|&vMAGĴ駏fQ)d8YdȓGFl}Ldc׬0vô+m(o@cnAӠ1Y.8x#0ث(6`05Z !!sQ"KlE$A>ҿ)yg%LΚb hj^8^I-2goVqq7h}8mjfnnVV+n1n¢'YDTc@LLT]P H=jE&Ԓd䢧0,c cg8d@-aA6 @mXOQ%2&@'i}'I+c"cNZG$e!y%*s3dpo~IN YJ*":Q!"l ʞʖ\ˢ*[/n xeyy,&Fq嶿{jޓ|O' HI\PG&8d%r"/oeSofyi}Iyx5e}G9 w"1}fѢz>̦o<=oߝ hX 7宻zL Ylk4u -joCxI?>S*aȭ'#Nȭܺc#bԭ# ڋ*254gO7ұzxzQF:vӶC \}n\\`jɖ\10A;SW5'%`H$L>hT~$@;ƭKLCS-LC膉,.P=\Yj)/Vkd&^07@4Խ?Ԗ+OW\8憈hυ]Ou[m2!epk),ZLG{=n>P6@߇ۋ3[z^ܛ6L[ G_ca~BxN7$V%ʖuJX&ACuwσt/%VCҬEm$ j]QIݴ⇠݆ҠFtv{FxjqomxvCB":Doqi7:"vJ.9M VƐ?)=rKcU/5+lzQmr qOW [6NJ6$&T5jbT1n$Ќa+!QG*u(c@[8c!f4#`ӺٖGa5骘[3Ly++$0Iןՙt9Xu2L#eu(h/ճOm~U tGQasy0a'P䣴n*oϗֲ :c)aّJSi)P@Wqgk#qcv:(׍A&SirN+l<7_@a h3quP*Ҁ(M &q+ѤbLDvԉVQCQunoVb9|!saY\NS(8]MOg^ïH˗iA>xouI03hEq :grr_ƣo~OTyab~P#0t8R8g6XXVUG Ab-a] Lٲ xrp\ 6p)s_Y$ӂYF:<Xl ),U}Smp{34wr׾='ǑO15漌.-KCrٱOV9wOk1!ۋY>WҺ\ܤѷҠw&eBT~1l,cLޓ粂)vqN79iN/Wkm0_FҢ1{:nd'ɓƺ`~?bb5~ L\R˫/[d.]/⫸!'A˅Otlc𼱫]F]os\qy %[{[rYpb!~ڜb%|dVIfd6JqXN 0mL=e1'(jdf,U9!a֡ }<*^FZiC2a͘,0Q GR1(;CWe "O/6~t?>ӛْE?'$lUKe4ձ ÑHx|tnf-+ ]k ү Bz7y|t_nhئFX[OOop|O7 ^􎼛\+Jq`1OWo/*`Y!PY?rzҗ-RJՄ) wSBB_ X4J\NTT< gHrɷ- YһH[,h5+lI ]< ؅0 %>^:K;v_ ֐XcO'llO2s"+aCkV"VrO)!D =T#3 @"t(Ltp)˒J YY6jJ쮫2.3@7k2D`|Yf^&DCM;Xk*&GFIt4MP$2)i]j)H ?8hnCMðTGm-|Ǫ:&v+ax_;q&Lz|K\^~~|xtV  /y"15Mt6Xt|O)?J4F\0k% )ж=OE!P,R4n FX d*a 2ly])jCr[ EI{)zQ/3o[3qH c2#}`=VtЛ-BDˀN?%}~oPИ\Ɠo2qo' }'gE>ϊ|b>u5!m5}Ȅ9*"ӄI\&OGCF^K"HsbIJ~Q`rT#we^@ɑA3euVwӬ?5֭t'buuq 0Fiى?fҘqVpnt#mӪ @N0.8U C]7E xY׫Ś`%2Q 0V$.c}<7)Ԁ1|#D )6 )S|LBC} 6e^yNiqkWPou;Q2U9N]m#7r+  D:UE ^p')ѯ󲧙][mins𠧻XY,V񩒐,1*e)1|/7F:b#A6GZc2nc4,AX);ݲD: ԛ{?/@Ę(cK ] IP?>U&-* V$"T:4cBе,@dt! cjIbI]1<],ϐk-ܔ=NԡQOG7fpTv{I<iޗH0e݀sԽ"\yF۟b{5zu9P4H*Br3G˅>(eM 4IEY6k`I`a)VXafƍRkj$QhXY`nTg&RC5k%}Q}^̋*r'ˍ,yA(K 83 mgSцʂtJRZAH"OInX&y|;H SA?Q};ͨ}BOZ Ʀ<4ڍrNHםh1<9Jd 4n^\YnӅ2ԽKt${<3% Ę$4>"OJ۴$ VNHB鰌M @$p t^<)vZ209*M9˛; OӨݭ=qww ֌ t{͓INP|8i3$1`I]F1ӑ7t1*+sZdJLH-arTۛ+uk78٦YȻiI`T&6/~t?ƍ`93C1L^FJǧM< ]'pFUe/RU+!ɬ`ќyU Z;K\^>GL&1 wUUJ6S":*^I]7͇wHx!j>7 mvw|r"Ew*Ҩ#m?=7( [~;]ȧFF!5xi{NN]pot:Ji:kV50bH.$X_~S.TW*)ɽKD|n4kX"&bYK6 > 1YKjBPH.`ZԚxA ֌j.XX钛 VcmK"*KmJSNK*bJ R62^I"$Tjdy\l{ {y9, 41PB#`j8cx/k3XRπʼn\SYrAl3œu4/i$ͬZo^3,IO{ZnS7$tAMNTdhpN؟&=p+o!~K`X9> +)ؠy1,fŐp )a?r$7FО&=ֈl| ǟNq1fnS v#glyh8{8M*=dVWT>7Q>%Lrt ݯDr}:1N *Nź.)}MT<~?g{NӜlS~RTX=Z9$mi;?1Cg;zoDUQ=8ʼ`T/ts%."xE@#ˣuU<^bjJ;c4@IOɵ*Ы,2muP^z+Z/u/A8;v5,l"YFnBؔnLǫr?+m|d&aS4>-C5:U>rWJ_xrVkVS}K` [xm10i] yF]s+ڕ3M~7o~A~g7hg)\G'݇]Y/m饿޵wewY}w"J[{G ))|\f-|gF3NVVcQ]V۲%rZZĄ͒޿3_Z$KJQBv t]YҖ\DXRsKU%,@ 5e#1nkV[|]dUhu[`c[*]h@D+mZ J$mJB-T|ϴZk2쐗zt7Mҷ֎!Þ׆`@~}>ӆlh>=Ǎs٬dWykYxU(F *7ѿ387ź< =2lIoI7-;ˏjOs-͂T f ǝ*UaZ)aEEͫo|9fj(NwZ0#-PءKq4 _TR웍X<}椏]o8W|Y(>*ܗp;7n]zPo<)+JHSăZ"W,V4&MK˼PFD(Q#pAsi y.dec-y(qu1-2S̨f%M 0.*gY0UYKhH2 lt[*)EG б`6L;=۝YMhNe^S8+SP"8n33t/SYmnJ3.}r85Tv#D$qIw;apJ67T>+6"dYLjN{%ı$\R:^Kq# Fma.H_[޺O!X)ҼX*Jz<χlcgÅ+|tݘ4֍IF+=Ec?NZx GX_Eoai=t.%zv1#x; ܭUQٿet~%V~ dZ+XkgY/᧮,=,Uʈ.*wE:af6T]WvzS?8*n.3eiG Me:es^MlMOO &>oP/ >F͢ƫ,dBDW `D |ե5Ju5ڳAW5 BT^4#9y>s Sѫ9k?c$NRǿ&2HهT>D7Ke;Ϙ_h>HCwI.6 OHW癨Õ?U=]g|:w&FK0D)BOF́oh I:o9Dw=~ Dw9D.*ƆF;nޘ<nk[. '-2i#T ; O j:{#J^ 3An3x.xv `kZQߋ3.5J_\G A#o0z<ڤo-L bLu`Y=,p8~?*f 80خWt6XVu>lyqFz?~yygiW]\jpc47s* jWty=EgkY2XEy+VZU-dJNZZU'^ϴP'fb̹Qoj?5;^59vVTõv=JGoF3vϧd2fc5 v5 hd$ WsϐɱuG yix5FBވKu"fUH;;Tp{||}T>O~q/Ņ?&í7Q=V 9u^w`ဌN/;!hpz_f3N(3~w\؀hFAc  ڏ>WiP>]V߃G F}V{]+y4j`PdO_~:`Ahr:G|krc(SyI:Xю53_r۹Rt1_*,r~V(O?eͼ|ܑ@o<'fQ@p)JlCV[Qrno.V@;_udawjӌ{VtL,PNZҐwQ:Hu޲n U Gn2QoX#]Sznŀ֭ y*S]auC Cn2QoX#FF'殭[ Oֆs=XDfʥ/*;':}iv~w]% G( .]^_MlMOO\>+?1y@q=Ƥ׽zpmtme7&v[ V1ؼQڽVAc: ٯ7jx+lΰ髱媘\zX !] Q:c$z+>B\p.d ѻtt]guf^֓R=i(t׈%Ϡ~%vQG@'`/Q~n0}6*CB1ً#q\!^Eb PIe +̟O0F`؝@)Kg:hX.$02cJbn (թ8kBf٦΄#le{cRE ixE>4KCFx*6XlV0łw49*tftfs8T-_}MoWa<~Z3Kk'k="_]weP53]1j>E[U?퓱d~Z]a>$vhWu,X6{8E$çƅ ,Zm9 !s\`=|T1`Sɰw.`*OT~m `m)mTtV+EGԍ@mri.\ݻ˫hS#f.Kp69su7Kol$wf3Se o5UNZƌ]w s" [C~Лxj{+b;٦$5vbnEɻӅ´mTZ4;J%r]r(3Z2tƤ5XMS?ՄkI29-?~~P>~nt63{i6z+^mi1D(,T,Girbe<%Y^ 5ҸcvN8Rřy$"1/f>uwQw.G+_14>&ͪd i)P$+S)As g:5)3u)KʵJ)dTY=D S^^}f5V1ec{vn#Mllq<4r.8Gio$zGBS1"׌t&#L(b ιbs/;S5GUy=hq=Yp$ yf\=;ȅ[kO#IB1 OC>s$PNӫ٥p W3a[1µC;%ۉ](S,| "(߽(E_/H1S"J1LBw]x? s&E=a\TtV̤nt;~t~pTL V%*+"QDQاq8|KeIM5#+nqC%^nF?mƲTgg(iyrL5gU]UU]%2'/("MqaFѠ$;- d0 MH5 Z9np:$3ۦHer Kȟ U×!_(/ yPkȣHk?hbF5z9Sqܳ69R;u4#9E52;, Z 2]383')EB9N/@d` uz-IA0"J;.վ$'7yB Q^ Ck{6U6_֚G헫) M\O'`Nh;󉼬&,7/ ,*ܔ;XS{\ zvq6nVў|\non0SŷQ=˛^߭>mddsD#6z˴5+1/o;I Uo2eMd6޾y`Gk^DPlc؋(X @244<5]Ě3o`&/X&/X)v* z.<|* BʢKso^ZnfOεF'dh5/<'IXI;PaFrk!C!z4P4*blďKCX:X@lV>ÙX6AjD05]x7v9Nm 6o[.Qڵq{LV m76(ڑ gz#˿s7sLRξAֺ SfmOE%[dcjلDQ .~čpT(H"m>q |{F ;C|1}toWxKWϑh^TQsQKǪitJ6WbtmW{ߐ@̸ a ^-a)К%罹UOÞJa!?Wی []} ν+nȄ8QTv${Jq%A(<^3BM*ĊV#n)mȝ0uzXE{UJQn' d&)ΜnO@A Xd\<i P;FKTlg*֗.CCr5,lvW VM3EF9Dl9c+_ 8$EfPJ$ X>ftW.F|eߒFdrKG%&]l'Ob,ǜ|y/1C-8Vx0Pۨ6< oh[u ~igsyqu:nr{mC)zvMyZ Ȇ&/'I}s̍$yfup7쳯_:È3qRhyTm-B'cbNiQZ\xh\cmNPAH i^@ ^~5m(2K;i*߅g<{&ɔ笌=o7ғK'5=E.FKlRB>,-.Ɍ:UAಗvCWp( zZކۛ)( }.I'D#AC@B2ZE P B t|Wt7Pv5Q HZ4;uXJ\ȸAPD ĽF{& ?oọʩmX_=%j?òplSnX)ADʒlv@OykWk w.L{ 筏s`1_vYeҫʮ'6{:[Ƀ`apōlBa"Hp9(j}n~zߌgmJa8miAV˟Taсw|-R#FOn!9uhz^K;OXD0]j\F*-pUaMysbJKW}՜Kǫ^Pu 4CL^D3KܤAP%oҿߦWZzs/U8*4 \u)p}bwMJ.֧YLwH_ 󸈬1w/'|D}I4Es TKf ÄF(#9Śj] aDnu`NLO[ɂ[b1/+n3E*`9IDZ z\|[Y>Hu~t@u|@S,SW~e(!5L02,J)\<(M,R $T3$}SrVwNv0d[ccw` #ȏĜ`ĂqT!,@ JB !z4Z+ĐAEXM-:AVcW4(3 tJ3Fը"K*5IC=5"4.Akq8K'3n=a]`);P1&jAI,x$xoQ V _"SE-vSjAd.dG:8Pby;, b Z5kfR* C'^ڌmV9 a\-.\Hl7,Ik%IeFFޛDJQ(XyNN#^vYIG+qC⋄\ 16uy*wZˈYwC_^}JOߊn-04b$*A5zt.'M8*Svao.N5nt.npZfwӛk.ۂ3}d32 TStwS4ѯ!N{C ׀a!Æʖ:8zM{d`&B"lOuy9j>Rnr_O/@HKўj΁%vV*RN D~ζ;\/o,l:|TKnw ^\t , ǟ+]R6=(DrS%Tl_nl&mi->P%=9_XYz?v).Rh7"[W rD;hy_ Ѵ[LLևELբ@qjzsjdNw4ngEY\uzn=H\D[Cs<7w^'y|˟z$ŎF=Ƿon~֋Y&)d߽߬1y23bÏ%faoj^墈]D˰xJ CqjmSK}xjM٫Z"B UXmWiI첅;;[3 0ȔSePbVS Kdpuh=sEs |:Ӡt۵ _j_Fӵ*oej07)Ϣئ͛ ptdžgQ&orh/Ay|=N$|g u D"A (A6hRS,x/, mrj/]_ԩKחoh<q9.S7rtVN(nfE|h&t Bq$fyΏsÂ7y `J8abHGALqޘpP&"M!NQ A&0#>5veNM@ƻ`y,ju\ZJڃ}7+^Y#bDVŧ{I.&7P}^8si"Gyv]$DbUS@I_qz/;[sPXߒ`I0O d*L#ExՃ+zfx~A07y&$cEߘI#C/`q"17kDR$7)!i\j.kksMA沽6H(]QII]njTLnXSe Q G*m]SL@5!p PyOa yY M \=k)VlY ͈*=q姲;L4 0Q>>F^|cٗ2~dփQG>u'RYHرT3&<)c{&=:O)cJNN.g[_xx;1csWqw}/#F˫G{f@MCpTcQ%IgR~U6bCDxmjC<2LMHѯgr^a9qI_fz)>ᗭrvLVE3/@ִ,iuԱ'A6u٠HJUe&2K$̄gVjaI0֖1VXN{1h#j_Iqo<6wu#S7s$Zu9ݣGE4S#v~WTk9īhdʺڤ5ZXķ`g5:ih)Ǣvis"WޟVw,Kszɲ!uXFX7}cيAv ~dhF>83dM -GSBM6y7)u!Z&\8d'o@*Ҫ&f4ء%;>k2?튑R\B5a&;\uNIHd,dpe8 Ú'Nk+<9w5!cHX/ 'UJ0>TaPQjTF{^2K"kzd(t庘&&A8ۺ}ZqZm]M8boii :%9ج B)f0\j4+Ui09g`njj-OnVF.)Us*<*FgcnBNksN8I{hgXoR?<5S9JG} D9AS:EHh_eSbltpJyf:T#ҏ+u/ '6 ;.^^K|.sӻw^⾐3N~\F f+L}կ;MDDHR6A__ߜt6_B~1F7~l =v/ H!%`< l'Džp1}D8덗Ry1cJp1x8R LYw/ﶗR1L?fI힄5"ԇvhzL`4S2tt:߁S6?ccOFY& ֞,<<bvp1b_|rq~#_(7퇳^:0MzOynlۇ9Ď/5LKnW~R)X˳{l\JCdOA"zXN2ԉXIDmH%?bH,9;4XH,xS=)-mZ:?S{-g%=KM<[ .4P ca59#o(wD+CݕAqoF5+]Њ$J}u hEHܝ{Ès /8u$KaΙˇo{ ڄkKL ERJʵ0"1 LIWc*0,K7GW6@3/`GcNy&>(Ŋ6Ngچ8x4R0 Jl`RQ!uLG+T bxsi֩nn/RL)&NCKjj]·%j)AqZJPKEǒOДde1̿<;K>1WZh~&j m 2c6܈W.$aԉ$p,;$;x y7 Gr$; \M0ast>`_ Dꀍ~08XNU@m5L^xp@JNd+{ΛPu4ޭՕ0R6x7, Vd,[{H\*.P/W0hغg|cbS`/u:@}]z7Az4gi3B. 0kbɔ3isdظ4Ŷ/y}l0JW%t+]z /b "y>fo~34iucpUXj91:klQQq9d4:p.23JWk)dc"cns *A$ndԵb$b$}E`H: v.㖰^J с@TP'^Rc^2F8܂QGyzyC}94ϊ[yW?];5JiIDo{lWp=C=ȕ!&ȝ }11drG$W331[OJݝuC:b楬| GB'(rik"?IV`l/.Ifvih!$-~ZĻGȖ\ =]h>vy~pq3^\jšDnGq;O! fBp}bom1މsGqH*͞CrJ}F{%(n/7u0[qmÇ5Em!lS Q}քB8sqZ5H_qZ@eX=/XNШ"CrW%kdQyKYä́>v`QAwлRxY];ik58 +"OFǃpsG$pY˦2" mp:Ӊ+2JeZnW-M]epPwExy:#pfa1r@u~WG\K ^i{>,Ґo\Ek'RoZ7U:8u+թ;FvD 0-"m'LFZ&4W:: =ysúbANՕAu;.Bf׸DZ&4Wѽu#J60{+:zg1鍧!h|Ύ*FM|9y)Dϗc1"CLpHXf4δTkYydM`"QZ#lTcu/XK1h u0978qiNAɧ ќ %DlxF~ګuj1*-{n(_%0zYt܃>YI,?ry)1*9 _kX 0y/lp3sXXJ=َ }L-P>&(Y؄ :R!i]@x GPrOEd#Lp] w8 {MHaYE-#_\a6d'%t=3I:E.I: *>}sYTi-s*X),#aXh\}55J0Qs|&[kT0$ATS$.1}s, W&!JJmaFʶowTCp$'B}jW̐@igsz߄&y:u~{ C7b1b,\).qlƎ2ޮ:4D7;P&u_ФO l?Qs vw[͊ TWٴ )GڧҞ֝1X C#F O8h-3 Y< X*/e\ *@㗩JYٵŹG,R/90-[+Y(dXnYU5_0u][<3Ψckt 2x$0V4^wRp(װ4T"'TcED$+TN{ ?uuڛ9ʧg@Ɉަ[  !8oiOE$C7T V$H$chlbcb-Պy\-`JG2Yl/ i֨ֈٗC"$I.Ih)qZS(h-۰Xvxޔ ^42o^+x>_,"gGsQcDSp%b  jSFAv91?b8/m1_/|?x^=r*CZEr*#&<4fST_SIg7SqZ )29՚SzZ#m)}y𤒖  'AK9  !D '|J%rH;&k!3Ό[ZFGrOI[OS}Nf[/ynURAr]>5+%R~|5QI7EF' 7sjX`fJ1&Mf:|.wj,eKW61WȕSwKd:K3(a>Ea!Ec8̱mwItT36ORuh͖/լS&R1KqEKYԝ=RmPht+Ɂ :KU uõ( eʖki:T!,>D:I'5\j8N}° 1YH#,hXei(65)&X KޓycIծ4"Ш4lx4hNO/!<|]|3[K^sɋuoa2%bGHrY1MhDclzWdRf6˰\Lj3g-C[’[IJsRbjtM_IӺ'bJ%g;:"TDnTlyQRl@"uq2v)!*n25ڤ`TR>&z4ŏkTlk[n#GloQ@[O(PM]LP$" ʢM=t[b8'L7.d&6Sp\ؚd+-5_.K7eQEݿ)%^zݛAD1i:@QQg\UM>~&qI /5emd zOg={-Uc`1~# =mr Qw7b`{=b\q_YAXs, D)F9fv֣ct%BݱvxiF.I+R_6?~sMwϓ *;PoT׳s?;'~v 9BIr >۽dnz4n q<r#MzFU^B}sb HΓ8y*Чy:gS8 U`=dRN%zc2iDÈl#Q}(qڌ.0ԮU|kh*(UtdkYZٰ4kE[eHX(MQ>kz[E2̪]t *D"ѷtMf҈XeVAH2N .dx$ETtj2+2A[3)-Mͫ)T nfMX)V6f԰43G)=LqCrN),TMSaNNX)}6_)D6G" z{oCi"s K-XR"]d6Ñ_@۾*Tݔ\<#e5уO}?tۛ4a#1ŢKkk=N?O}\K^ wúz"p;\3BOX>wƴCcv8˞vFQ#wT7w8J4I[/V|ۗem[RVs2lk,XO@'g'2i OCe)*De Qt1˸;!nXBxt^"ݓ(Tc"Dkz4_JxL$ LYe.p1J#{GHRU(~ΥqIYJ91Ր9*K3ΨM^4gN_+QLJB< ([;9O!!z;ҙ~*4eWiK\W vt=f+_ԫu~ښ$ m0DCt[4D7dcy 5J NkjQ! s͔1H9JvIMcB[No7c~l\m \Cy ( g[O2 Ʀv[!"[R$*VMsĘ+Of*43H#s,*UI  U<! -ZWO--aA FDA πU !?.˱ɽzD{| z{};hӋC:'؎1!jY@R(a@`5qLsfc{ڿ K]wg;'M?.YJt!slD#G$ƌ6{q7~gm|?|NR)I,-/\arfVz7 B4I)2a@HبT+i5R!p:VYK|JjQ ˣe:e'(_x3wj3>}?=߆ O07/=}?={?w="K};ճx;ι܀ݣwzaX6Wʺa~ۢfl<{9{ύ&kP"Q$NG^2ѿ53|i7c;r a:U"EIX}87k =!;͌+o#[2~ς3v㛫ڛJ~w* k)M)9#!_m[f<ٱo{4*F{{W1+p=U9hyլ=6z ld^Wa+y:=YIzS'i 2vh>4faZ%Y-FC0`=[^.m*쬩ڹ9;qʥ,BxNݠV$w$XVH.pts_ʒP~* ) dZVVSe Y)25Hq"*dUJaVmS6uƜv&#@(֐Zwzi(<כ~ Qń4Bԥaə'T\a BXTL,}QDr-uM_O@12'܊]m -O:g,ӧl5]qS7Y'fNFبdPԓ]>d^- ; Q{W˶:q G{xH-.AgH#)Z*q/7R+AEKZKY!aސH!BKYK-̥笥 B1Zyg3Q_+<L4NOPpe"?H7;K{-5YR#@_Q4_]xE$ Tl CT+J(ܛQAbRr4 nVljx2F' M(%J`F1kZUu/˭$ӣlZ oU9 =fץU5I;z$/d!QKAM7PVQkUnuJs(lwӱݒ*hQؕ-EtY-F#QإH}V{I=Ω=$LcA|e}sJ5 IFɆdCȆɦHF^VOtNߺy, JTuP J9B Jj_i*r8IT3ZD@@UHU&{u*^W%hW;oV7[Ka/2#4O#\F AiM#*5u.2)BO`L!VVo]ӡ4Z"#%GTq/I}-oH\:ZEF^3)$ Q>uY_1yu7[*QɫKfZ$}jP,\)5rRd+]t]€Q;(1i.uZH}X@o9cfĴ2BX})mdˬ`bJ D q$~}DKJ ڄzݏպ؄@nԛe%XO^jϓyE]_myxء7|M+Hʈ'1,$)AH! 2H57 s͐Q(ZI!sN1#$SoIhʹᙤigV&U<ؐTk! OC6T|UtiObU U~~z5NJ7p̅'c5_vj~߇! x`?vl[]-KHD?-#!">G)mpDб-iW+0w١gw_wKUQUϿ_]{I*?fov t}ӳwmyfvt:Zje <7:zmaK6U$GDdFF5ݗ9B"ҳRq4ϑpq݋;l$.KB[KI\pٳU&Z V.&$Q2vbyHkI',,$ 4!"%Im(!?ulpn'k+IM//nsҔ#F {ksQD`e'QtYO3z?ab]%R <'IEi\1].f zx a;@8N:8)+N$ =$"ўAQq>++ 4sƬt_M 9v ҘQGrQkSd)wv;{Q[jךJ($m4BUSOy;uR~멉o=yuSլ!&%H'ۛgS GE\ͳЧ⊰ˑ'0(t>{~&/؉o4y F4Z@Ȇ>aϛ[ .5wvr0h9*Qku35PCQ.y4m9Smvlǰ߆a(ww@]xɧ^s;8|M~+?Y/ߤ&'CN…-ë!+\9(~vCV@ֲYɥRusV"Tޣ ̒7gi])Bl0v'9H?u4mJ0;79>d4}F2Ҙ3ڞ!lJdd,Q%Ճ^k3WmR<zG\;a\|l j RcF-~g7Vri S]fL5E D9: ʾS5fĂHt}#$.Jʮ;T+GU#NDmW?vRÁxZ[.L`]$^;wK$05XؒTrD:YR幠UE.4OC)#g u!_u*\c.ޅ,aH=lqU Ƈ30NY_txV{sV&ŎHӬAu{}[J^Z=3x DU>':Tp?} *rBXZ\JMgl>dV"1.eeU&ZRYc=.}! =Jg:&nS 1\G y Êʧ(y: kԥ]_&VH.bdM̡1d̍? !^1]ILY7 y#9n7y>:yb ޳+#ӹx2vB{ϐe6x/:˦>F&o͛Vϐ 1閞+ˆC YCp3d4 aA5pDQd?jsHn)e&!>\Fz@dGUF8fѿ42]@ ^p1UC|6Բb8(d'xEh' _kIE3Cyjqu!J/6oɢA| w8;sE7{uI^.~Qƅ3@_1^/6c pLbAA)׭[]PYFqݚYkJ ]m>%U[?Xu~\Hle0wWrcNIah%#P֏^tx8ϫvkۭlv[^C8N8T(E2"!?#6*D jKXː2D)uйϧih;paՆQv:5ZmvnaTVz۱ \)UA _=qk?>cZҏ!H@#"L&F4*&$,DhQSntkGcnlU|Gӻt}Tmԟư6d~*]*} |fLҟƮ"IBx0V|^Yf_w?.TFxh"χ#b bWFp9Co*{J"َ֯kHf2t=A+P*|PW/c ;^o` N׫'x ӟ[Q`΍Lw"H*/Gpl]O__2(l˂M~\lLA,¾vey_әkp ykdɬE\i}9[nm=BbiAq1Ja'XMo?Fsk]5nM%Fj-쉹I'zRAc z ^iUͫW؄({Iߡlخ \[O?@I퀥jH|KGӗQz3~DYo^ 'ϯg$r"@NNȾѮW߾[IB!甓iOUa'7P}'EtϯA̟O o&Oe ?oMqc.Aqg+CUtej)?mpY\,UQܶ@>GT_%@ &53JE"-i:x%H3Au@B*8J5oȒS }܃- a)1o: $t)iWv949@ fIǘiac>BrvrN"ayP˕9Kv֣ÂN5A nqYԒi&n$<n. Gy.cj_Ø#%+fLi2i.w EAPZK{M_*^9cRtSfTnKcjѩkJՌZR(! 2ڳ:Y$_,TĎSQDZ,UæP[-"pK3J6"I#T]B缝 Ece=[wBv]<E Fc0Ŝ($dž}3m tE{5 "cP2"0$I"C- DGa(:%~a!=*3=l4> ?>ڐ0&4"X8"7Tj&Y|\9Fl0r<%1F0c!a"2 fq(1 (ϐ WaI k7wڊQ2 :*vjj_Nw 34=@z4֧zֆ8wPmǑ&zO}&ԍ|Th%fQ jD;9#[s9>ϥn>+MdR}-Z`>`)}bmzQ %=ĤVö{@ zY(r$3uNY8k+Mhm fv. OߔтɃ;7wd1O:<·T[t7i?ꎢu`Qxk'wiiY'kym11uVa6s=OZZڐ˩tuٔ[vܽũ!M\Z=~ܥ 4Vt줔nɨIGaQr\\%IՕZق9OV%FѤ28˂A;yEuyE94N6PoAqXSaUf@%pGe0{_NU/(@pu}A解X#ىnrFP3R!GA k8zuUu׋0 C-2jx%hH/*G Zŵx52g\Uu5{=ί^N Nܱ~5 RW;u1 RX'iKqk]}Z]W%Jm2E*oP[T9f M$u#Z@ݡoMMm$ E@$oi7n8v+A蔾vzGD]+s8{.r @cWWlN!>+zb9в>v0EɅl.y&)^ S탡Dhg+@V KZ)}E[ i~(>ggc)E*Xmd-.7xfe\Ѳ2UB;2 1H.*XR57 ]F}I|ăshkRY@XL㚛4 A @ ƃ[ VD1KіfҖ.0C$=nGLN5ϳM~ ɥ&HbE'S/Jwbp!:I1ˣZ')pD+ |ZےŢx}4Q7}y#W =h^2yMxGJ>|`̇ȣՊp?G[݋'c${tHχW!Ziy3ziyƙa3 o9%9z;q;@5#?eS<5ܠ@L@`6HZ΢ݣðJg ?IAVbbq`cx.(R e@*.AƟ&3ښmhOv݇hz~<O.&jۯ:tTG?ԍF,?>뀙Fσx^2ԃ k\֙hh,&؋h""=nC&aA4bw^$P>;eG"c7@@[ptƌ3J*)J;THJ[BFU?O4ZĎ'csvJ.R/H<(YBOEZ?~G\Kl};>"U0 NӨT7+G;J1Ҥo{ H9<=7?JUt 'LJFfVBlzTE2Qu= (.Ip\*w[n]Al3kzyݎLv^GH?12ɼBX 9Fv3P0Ӑ@(1h{LBBjieծ4PrZ1Σ@CTvѹK+jnxs^7dR bB~CG uC#G cD/ Tolpܲ`IX 疭E}#$޷/wŻ'V0ϭ_¬w»W?dn1qF&=\f{4ӓsq;{4v`ww'<79L#b{3( D1A&qJ"Fw9&uF9ܬVED"df> L@ 1.0#&VGmqҰKګ`1PS9qOyTeg4 0B@CyuhZ%nV& Z00i}x%b{W1LJEq`9D"?BM⑰V"rꢅUdUJq//Nlė~<'8͏oW/]r;@'M)eԿ(}ǝ}Z]Wќ\q<>~Tӣ8ӿזNo6BS#=6iIIBr-VC]Fn]ͨJi#:"OY+wڭ>S!!_Hr=q^}FVJ)}G6)"uoVZ)S!!_)%ebG.Q{巂 HS|a1` 0_qgPU\"R1KL% "A9=["w+9+#/^)x4^5`J(#c4 xJJzNi͗K}k Yj VUk>ԤZߴ4#-@fgKwRS8)Qff X"s1P -?, +wL>=HH16[@L9/r#24.r nRw(`r YRXy qkI9U\M yb: (b͖.XO)n7OAbЎw<}f )CM8rgH"m\-MnP/nJ.(92AyxsG "` G$f>LprZrT'H-){qᴖT(GWbI`b!pnֻR9ÞfF(&%$d?Eޛ^C&C& $H' ` ) q#'q.QݽNP Zи[Wzbw^$PHӷK$!^D/ d^kfJp:#Sf{Huav;M~v26<ư/D[66R/lGE;{ {kмNcE\ jrtA<+EHAEXKoڨ riyV+8Iϭ@mHU_?_ //i5(t7xT8J^.Y)FΡ!8kCj;+923uQ|i0 -`o?\}tcJݰU>n70wEfmo~zg]ջgGk5x\:s5=\|C&9| 5Aa K4~l1W7~mқ ̜g3,:UʧvTdG4y1,1vjqSh訟\覗nPfry!Ⱥ,WΒ&=M cab/AU2|GWT&*ҹqNC]dbQEsVQD/̵%YcHT(*hTE| r$i]c7r9\*ʶA=\6g3VͩKǪ9[Wc@ڸi!׹<W5<d}۪>ë$8ZW:AǮ,J_B}FM^VRyVN$+՚g5j(+=g+dzRH1 eRp<+Q;.{3R%xTDcDw;FiQ*w4 MVם"pY7f@OFZeg VKGum."C+pl2OǕ+g>̗?w|Z$~X2 ȗQerVyrEv{]*&9y_ii p R.nu t}"(([UD)M&GpN^A}FML44J*NEwI& ݀9а?gn1v m;vEd b&`BađSorDA"Q,ey2,^NPMeޮpo3', CPyQ"%:DcEJ@Z%s 7EH;Q;mpӱ:]*óR!;XY[v<+ՉeEXv<+Qj8ZFw8RdRtyr$vOc]g凟ʏM2g46wx~gIN7΁d ~ٜV@F;O`5G{㵳LJ߳k-R9,T)pKY JlOW*ָD 椖FXX?g7^Qq:E}j3ibq;/>&NIJ8,VTԸZxץ7B㰚/^ؼݿ&tfN濬Lm8:P4~6i Y7|d4Z !rM`r24Th+%*H K_TA*z8 G$33˩b UNPQsMlU('B PU^JD)Kc+O>W!e \x0dh4z[>5Q?\}œ>|#SW_((-C?cvoM[Q*_DMY8FB%seTZB J2Hʮ%[ōk#&v~v⹻mU6n{Q6RF%+k:r͢ZAIA@>qbhyBE2=` lJAN#CM &k +I(]7;n]fv1V@aHn]EO̊Ruʱn֌.φQKQ;ٞiwi܍gU7) ΢uy\}3tA}o3?O0b ڲhXw|]Г{Y ?QnxHPMI ,׏!NLJo/_qğz^#?\m]F!PnWr34<]M@L^ Vv@ybQd()_Gx(ӕ:y{E1,V5 F C,tVB)j'oHc^1k-x`X)'u=p!9MuzW8'v~.rUmϓqo믿|N ߪ"/Tv*ͼ0tkpEϻ~a?@FD>hbn.cyBEka' WpG+=?[X*|kiSͮyzr-1J 'Y?3ptqRép@™ĿLRyVpt|;RFKAOyf뫊M{Χ{Osw+?ՕO˰})zq{n\WR]Xcbii#-h:2{y.V{SZc^_Yą^=wᡡMHlkwkvIѻ t~w/:S{yL6|&:Ȧy7-ʀ0Fm-Oԩ0]68`!_lJ1mJPhX?b7,/9'^#FM/f"3SNYcad]4ʍCA;,-1r6<ܦ:&ܹB|yOڋ(R€Wy!diJv*B1Mp< VN;*7Z/i%0n̚Ƚᷔ^{ X y!RI7'wEbemo~zge]Wﲟ ҟawwqW~*gMh)>iФ}بXY +ugJS dɀϜ>CBz%j5i6yjND՝/:#XMX"Bp_'1>Lc>owY%% 0AU C2djZ*FCWleRdQ_xp-`hM h]$cо0-A[)?gqJqxN1VVic3cp&oB ,riI1yET8T r(JK{ԻAg H'!6Ύ2h@eW5d On\圗S`2^k2o~N:FtZɶ 09\y RE!\W6S•+!/9o8kەT)r~"*Ch % [YIٜ 1vPkf}܊JٻUэ-ڗf%݂NJUt015x_y`ͯud&q91g =?; 23GRϊ(A>~v=xnq ա-7 PME Iv{Nc?upR3?moR>zÝYrG9٤DSj+5)!F8vN$ZaQjc\OPNN` /G;Դ%{&:_pOGT+!@ vh5J¾o's$=R7πZ) zjsA]0;h6u BNIͬHasKu]n2/7~]>}v(pP1(= Y͂Fc KvWM[]=a@(?qx>.QrwO?ib9~e-p׮9גI뙵kLn۱4Cx zv+jًc]`vZߝ#V(4 4ʰ(l40LÄeEEۘbr,jHsqs7sסM E`)tɠ${F-xJPCf 9Ng85gb%ƛMXmr1\)Rs5 5cjڎ߽|zzyj6╫q?؛{/Otd)|AS=%3B/ԜML݀ 4v#$H!I$7ԺPp(tRXo4Q#&8AUKkXx"'ttڡ/4v{7'Zi%8Gqw@F@TѢ1hMOSRjx+a =h6ßvSu joR{q^knzªqxpʔ[]`0\BA$Q,㡟,j*R(G5aQ"xn!i4AAYd;Ja/kDR\fA(L2Y,1AD4Qu{p~j׍ZmVFea !#4tH8œ)4RAS$kxCӌ@ 5Z *.K56 fYc ̾qj͸tjjjQ\nh%,1ֽi^1FB(nÈ\9-Ȃ4 {z=v1Nndfk|"w3ɵ䓑(,2r~E]:}x_|ybSgo!>Tjp 5_G$J\HUq{ hBZV [֗V$My!$ݙgMsVhR*wf95Z&+]']̯G*g6 Ut$ыhg5&%!35ޘ8t"'{xpޭurx`$oC Q_rCf& Y0oϕO,q|w|u|2Q]WKiu;TߗR`;^*خ>כF:Lr:%/5zہmU. R1C#IO.Cd9k\=;M3A A }Gv:/q⩡ij1$'!2ʼn>{'ƮvS()m4:v2gnCͶF\DeJ]rQZXU!dtQ3$~|OiǨHy]r{ņ.Gxt.o98K;f"ADr|je<.mwhLC1N0f$'#%$P:T2rd´><| t*RFJ{,=O-MP"v̻N\[e\UHhd"P/cE^ CͱzH |*$vRDpۻDR*{4KXGܗB^P*+1>%kZ ab! ?8W% Zxx;ÃRw"J%h(.HnEI%>ͅTfrҨj#JNӜc#%'*g4FC#7Fh+T@2B(!)cTCȥlXndfu_z _h9b@*pIt~~`^:7?0%pZ씄+@ DA 2X1RPzzAD%XKGM޽jJ?U^􇩅Pȗǥ .(`x0P;ۛ\uI>S~9hJ2Thu- 7hPs}:[/nb AF]O[i\Ns˼H~A +r)E qN ҙ iD^͟b3t54nxhܢ103M޴nP*!߸p ,McfßrLyo}iDR$HOQE鄔:Se\z\ڃ^_d "HѰKKʭ7JO8p0k;wAxP/TC%BB^󙰅wj \'0P='I(C QfR0 TJ]YMai/S(ٕ4ݴR;bTFul$$QwI҂2ӗ/oNDFM"&L5Rӌq\g?۩OI)ݱ]5t 9dvBp`F8xiJXe$TK>rY{x=ge흐J-&8X,'Ah(cڎ;k7!CΚ~՘k6dpa)`?QX1m PRH9MuaiI PZ8/cO5ѮqY{u0jŏ5pt̓3Z7ilO7+%^Mv}#\0$b]c][ |x'^Nm+A376xxWcm @I̕n+)a*E Jm8:u=(:-҂B&j\Q p "Q3dI`6Z|u^ 0ݛ40j (%Yc'߻(9H6ϑd \[z,VYi?cn~3u;yLJtZCj it}x<;ܬ>O6\ϻdK?k&rJ%^\t ǃi=H yD}^5mE[A‘a0+ \@]runjNҼMP6C7\(kOdeKam2A8`+̹ jJi Y'LB(G/?0IEnW~ jXb&ʦ@m9:%*>&t .ޓ`&CT-&{v b, x,&OYLٻPL0xo Kdu/n$8n˾|9R'Y{aFC`; "K|}ݳm}H8d7lAk3{f5e^kM_>IeEs÷O*UQtq~؅8kuDNf#t\pe87'l OO?n1m3du^ODPol{ޙm ł0#U q5Pi[guAaщ%#(S6ٽ0XC{uiď9$žG_]@W!=ۺ )#g+\PZq?,z2gnP-d+a>mFD$E-tϿa[l5XM)B^s!: p\9;z8z8r2g8q\E(Gmm57Eh5!d$P ?M:e2KOӕTY+y6_o8_`РdHuyD"4$ȿu^14Hq O4'ݓ~ËC;&YNӪF7Ӎ؅$r˜EE΍,̃+a2csy`۔0 H;%}lhZ!MݎE%:`HKdV$ylTA+k/BcQjj"Ky7VIڢI2\$IVT̢ffkF]KKhluW׌~L)8ַz,\. Ă6EZqd vы}x: ;wmIā4Mg&{ݤ\'_TmiD杴dtFs_u/B~Y4okOd㑈LVܓP/)X(^3sLQИFCwR(}:\>2MGW-nhP@vQ=$P֜+xhjmէrbbM1E\"rS&cNBW¹?`Nż\Ma}8aQ>ȸ y! 8cdyyBrx 0Co|Pϼ^_+~&c>p%9u=u/m$ZkqI qFlXB%erPf)B YMD1z2B_ 87ǂh z5WQO$XwpcivI\jV$NQ?n$r&rŤM)yX8d_|V]f: _&u^_/lDa,S9{@(IlMJcưTuXbO~ȳrn%ĬȀ,4h2oletRAN:Aa@Z0lj*3'1T8 w;<,56GHA r sFL!5I*Al YiǃZJGtkty[E%fr&[-x 9lEQ2F{}m/Zݘ<u_LwHٺHO'd>)|=qϒcnhﳃqO xhY0qMqqP`e%G ހC+otfp5h*PBSD"Ĩu̾0NZG"E7QQH(aI8mQÍ>t w@FѨXVvOϼTԑb\&ς&$ѳ^>!CG*GOLbX'_ۚWX/$dvvtuu=eY06vBUWhvpNBqGH[ d2 kp*q5gj1Yfw0rHij[ݾBKtB>8RXfLtu1GcJ(е(~yhy5,חYu 1\_~ɋLRM Zu'O&VOr‚cƺ]( $bq4|W)⸼Y˻۱حg%% =#`0a*/г:@΢vg}YNSꋎ<\8POGJ>OؔD4C&%55%5ٛ(8Z~J('RB&V:ZN Q_dqޢzL7{xb?ƦLZ]Hbx_T^fas 0o$jNlED2̳vnǓ䩾j;3\U\&(&p %;΂ѐD\ӧz> \GIǝ^bcvh/x RmӅy%SE&T= 7ĥazg'd(gV 䒻lU{Y&}1n6pAdSCsQ~F?8`$[脉*jfi,($2\.[s`y=1Ĥ@9FZa3a>' 'A2p8,x>Z Ι>TuX\!ъb(.(s1\Ե¡W8j/MHi%~DZ 1K'3I"ZsYꁎ\%uy?mj}jey3kκϗc)GDf\;VXt Q{ĉoVl~;۔7kcSRJ"hfĺK&R;V)ZH6F =}nZibMҚ>M&i͕gS&KɁ,*$IqKCg pZ^k5aF핖KBV/"ɌT6XDdIz/$!=v|? 2́eYF'-3ki{"c&Mv㢡^v/\n' bZ˂r!mK_zl5^fXsV^>3p/c;0"udҊ]Mz//zc7JofWC;_jPA׍77i@C՞(/on37 ȝRGjԤoǀ-Pdφ C*e;%~jwe+-&˄B{wK\̠b5qeJ eYR_[t$;s N9Cw\c@rKޣ'ڜՂЦɲQ? {EmW` zVt>{ 0+QU+ov;˛K[a^ ʏK#ܹ3LߴošdCP]g:Nku'{ j5>[iPg( ˯us8.Zb_4փgw4׫_.AjSH EN,$Q{SFE[Kb9t5-yfiO(=SRkFob‘zZ>lXa5xRsW3}I_maW5i+ɮ+Q^+qegWr;PBr'V/~: A~Oˏ165#lt&6ncfv?F^S6e~~ S\@L31)zR'SR*ez9,$*ea;Ip8)R˫~?_'z@bk1:aDwno1UE|rFAEca f.z&vY,6)x:u'e|; wq/oen_Jv/7iYI|2u#IQp^ZBl/%&K* !},Ec) {blY>PRJ6R<ȃt:bc{L_ۙ1mz?Ic) #Oϱy(ɄyF4n1s*,OFGY0ic0 &D| :X MBd âQ R#[kl&]98.j<=`]\~öS:9 KSS׏m;5 BU"B/'S0Eea7ooO |\Ml{r@*0yoL) ɝ vjoC cz\N4'@2eUV=Caq=6d(!'vZ7F5#E;IK%$\Q$[IjڔODdě 7[7C753m Fm# 1"IJXF@YK,/ɄH`IS9f&T$m)G sJF=.٪VҸJԤ\D,NV*$*T> ^[ۛm&n6-6}d2=Z[_Nӓ<`/`6yS:Mt}U._ (6|8TKj TnO7M#T#WuF*׌ I#E3xp| ֔=~Îl1bz<-.<4][]kn)wrbxr̹22] o4KxE% $|#210scA 䮔u^15**&GS&= `;8@W>N#U<,ֻ :C~q;R7-nBHԳ'ٜh-M:PJ̟-^'&>1uw&^;OIzGK]#n;d{ҲO,Ӳ맟}wwXi#* kyjZY׃ -qxnYhoѢ͹gw܃s>vmU_qg~vrBbJ/v:sEp:Ȼ !'7p? ?62lO\Q֚N_i6[?콿F"0]z\ɽޑƲkyzPhs¯J&#?؜$ T.:=>02it: !.W8ADeaH]8+0 4g@&! g<2(vwna<@ J1Bb0؈[ մ4b0F`Q5-wM\Ic}!~D]%mٽYIj2.K{H h #NK$h1P&`cmU%p|zɃF9<8_k ß ALy"T0}ˍT1jL_.b/Nj 3hB<8!m=FQx)%fcıSɏ}s|{+J'QD+ # Qd(FP09|@Y/RA9 T`,@UpSYQQ M9KŴ7:0fGKQs f89ru~x8kMM_9݋=ΠKuYq6*صC:Ô& FcoE!&Z yXGܝH)pCjYI ]!%q>^$P{.>Z:9|<ȋ ɧCL C:U[gM`GCk;:jb~| =WP$. )9NN\| 8zz3ze\'582%=;,D3!Q#.D` &DIcWG~WI'~!]^"FS^'5ׂH$˴® U77K{5q|U P&P$p+1Ҕ Dj#TZfEb kV2;ݯ x zR!3T{ $0%2+eK,_k. Ff_3na?=4gVuϥ>#LB}"IC+ՠ#-8¹RE'Ή⮀%e{|H#Ap /M`4㈋(AՉFX*c8+ฦ7[o^'Bro̍]r^̬"XB02!>D"ќKH՛ës%bٌK퍁h6@|Y8i"%Gz",WDЯo7ƙ <4e!RO sK -{G=vYnooSNkq@ Hd",(` ?Y98X>BaaCKndSQp&\ "PkO*Kűnn]̎ރ%w:)|T,MG%O/Cp !5Ya$İDiqwA \@8 ~`C RẉݾRg! -RV @q|A@)Eu1`Ns 'e}E@_a= TB l$HP{m Ðy=zᮢX#{i|# 0Gk ylLHW] pA;D $dq`.ȩb|w5jy.7?7Q意r;ϋ/H -O&8,|+ LQQ>7s 0 o+`8~{;)?7He,)u9!WH̑øF Pfz9lvA 7qh:p~%&  rFډvM4 ^^-|깘M#|VC#Ɨ^2 vLK" r*hwc]3` lː$ mdIJùG.ɼ='읷vF>bo'L h@y!148:|#2Њb`(E2jEsB:#Ӎ׊e._i ,"<ԊZūFE j!26UI %m+ r ?7.bB \hC`&.I j Enˆ9 'fMuJQ~q:RQf Ԥ $2T(0 %MM^*N3r 7g)K `-CL M,%*f'쵔'o!^kPûw $l0XoO7吖ny]Z}7,|FX2z"C;\ }f>YGO üTD6!^36P(᥮0|Tca{8  d44Q9,ȳeI>zRPP̳geT@ɕ>9Cpڞ6'{>E0/SJu7rM6iLX@/9z0Pd6tׄjW Fw G?.W,{&k4y)!.񬹰nzJ-)~hLrGV!0Ӡlc4X<0L #9`τ( pD#MZNz'*~Y=}Z0+q)I<Ks Ӑ_'LJ!r %.G)q<.;vt>!ɐW+T*bBsV&~}T]i X2S,W(^6 G_?UBPJة:~h*ҥbFb.[Q^AdP(-5))OǬ_$JH[6.TI+$S^yH;ЫI"wR;ZfNX c4PdD+2 -Ȓ$u6U_[k9G.,cF.|]V IP^\I价er|{ݢ] ߿[[(:ZFmf??^_1 u>ml:[@a>@RX}3 x1)HI tL/$JE^lB0 ڿG#X%$䡜VOB v;?GtiުR*(XmC~B:̨BYˌd 1iDA‡_Ą+W~DV<8vhұ7b4|qq:.4..e/hY燤}R:^iin_ai}RLj~eJvsHXfdZtRFzX& ?鴆*BҋLzwȋw'pjSk"aHy;ЖOAy Z'@kc"RH^fC%ÐWJD*/z!U*ZEb&p%:aNF*$H;JHI!(A-ܕ3뵏 S?6%YoȊ|Pm\)n7cn?ᢦ,Z҄e'Pt2(_(O uv_Z2s;gn%:M .H4 I&P\n\iʃR!!1/`z=my{`z0:\!=2ױ\_ ^$Ny(Hv5EX Kl)u!N/qA.C=.}D^p+*> 0A\o>GЮZ-CヿR쾣;_|~`;1 fx5zu`cd㲨Bx葕X AX^A>bkX⺬o&?bס34-L̴'jufDv &}s+FD`܈x/lbyQ#3K{sM=9JtP#PGOXf&9*5&8vm/74ZjS=bH/߄mǫiRZT:o$5'~M`ěuh-wϹl_jN{ɦJZR $Ëb+v}&ti춛ML!嚐5v{<Ϝw0S7 -UV )>cE&<iφ&3lE< $y9:rϏ/g{ W S;+AY!E.UȜL 9Yts:ܭ &ƆϨpA1B;n-x+wu"cP 6UvܮKTQTSeR\Jq FZ\޹CkaA٧^uFLs}jX+KBKaHռ|{55]qYON)g "˝94PIdVydn&yy?'/͇jh>D΅aC8]]`'V;r,=:ޒrEnₚ$|o+.ee@{<ʒʛO8M,#IrO2n2 ]#:M ![ $rj~ɒ h08h;`K׀}y[4Ku=*% eoJ^z^׎_t =Hyi.5H^2 R ?Ӈti{ғ!6vO|)b Gn+^dk7ܽ>Ngv؛k2,6-\v/0,/G:VuӼNze^so۴y`y#H rG,eq^JYF)%oH z{)Sq^BR$ TW/(/% 9ͯ! R xnWuz2}10 f+F S[O!g>:n7oON x:-$gsТh#飝ut2oL簨W˩aYR'R_ H<9$]m@iI*:.leT@ɕ>m0-*|⍲v%P(X5DfEŜoaA=bMx(,$T6E/gӺgFQsxXaW, { њ̏TtRMciYbĥ3;kŸ1A*.ӕoKO$K45P[m_Tۛ((zZuXH Y2>a/'6Bq؞H wʠׯ6ǰ6ς|Fv߾;|ﺢu@,8BCKc%k_B))(d >86nWtqsq2 -Α$Ev d%?zň btZJ  F#,lܱoKR7*O"%`qfiG4QtH_eBt!ހHAOY䈜1;ts(bZ}GP v\cqFP"jYe _Y̲ kmm+ն8w-~ڪ]L n2c5w5.' v"(V#]k-F:3ڥ0`u/bٙ$D,FL( |IAZ^:'yKx7=5c̱! ˤt+ԫ!Oŗ/GL. ;0ﭐrB{vϡEC< 19įFB,OWdtk/a~s'4`/`AۅĄYqѥ݀>e-%:)Z‘7G|,|;! Ө.rT .[OdōO/KA)tmI0O$#R™ BI }pMaU0x~IyªF$2/:nXt46+=% :Ȅ]J1 5e L G–V`L8^STI(KRreåfV bDiUYrIcaŔ ܠ'V8Ό胼&:u'@tAj(,S0mB&RhAc* NBKd>lŌAC̀]I4f%03 b8nNBu`>'Uaٚ'۫72IrlY~^4$&Aub؈ِ&)j m%{4.&|8iEZ/S+4 /iH2$aJpwdiϠR$w 29\oKV y֗rؔ)m *)NX҆rL]Ԍ`•"~,'X_$CJ1]о[8Uw!x\N;x}bڻ埐%z.,;7ѭmI!9k-Lq_ gτ RomBKJM,F,br8Q먾hPB,~H=)d2ͩ< +]GEjJJb@J𦟨V JJ9Jq2Z:WSmETP]v˱5Yqi2mPI՘Ĺ`oABV[DY)їAX:/T+@4+e(Rt{T_4Va9Yq[)$Z) JюouÍȻzޟb\\=}x)z=MJAyo %sPy;j3 r[;\H<'@Wwn/K/Do3b^Lꗂ5\}Y}u'KVѣݷ|j+(e'O2Uwx&AڣuP`K2W˔=NBؿYy'en͇EwKFMb!GhɵBsaB &Y@+D! @Q{S%rYjRnXܲCOwW$)g;^Qbc}F:8ў8ǭ`X_) lW@J o>5Iꛥ$uZh@ ĎOq6b;,nCB(ft&K*Y'e)Ԅj4H:`V!#Oxa%%1iʁ Jն"`˖mر-IpasE5G T+Z`T2BUkĕ;h$v F tPzi pk7CGW_|O<{JxQ烿 k}UKx3~7sdD20-= cE32S:n"0,v55X %)8hT[]E(w 7.8CST*,Ah\p# O)"h,Q\F`sdipyCLn?_NF:NʃĹ嫳{bݭ2ʻ9PHL8gO_>?)aQx<}ě~x!??,x2SieߙzUL5_j CB 7z@gsrwDa"C֖W[=8u58_u{W\eȞs%uKE"bĎ|{h7Zj`_gT}6>1mj8%^J.v1e"ZZsɬg N1)!eP̓^"=ՇO3l: pC"C@3Xozxj;_GilO4x՞rNTiK".$93P]*3Q"ŅHK΄C"C7µ9y$MW(~̠$ 4@;Hy3 ;-o-@0[YTN5Ku=aʂY / Ҫ2XR}Kx!%S̈́R=&iA,iwT-;lOvrI57#u RszkbԩX} ]Xwn96v(Fn@jG{n2H1wn{N{n=[hM5|Jq-W)6mP?1ѻua!߹.N4hl6\JKK.EjQ 173zAk*(;hbW8np󑎺>'ŕ|'oīoqx&6wH%ca|Ro:R-Zfn2(vޑ:r7o{'=__N+AQSJR0ګRڳR%C*i-i֡p;-%5]\h6Bǹ% T-uP`K2WʔCv(5fB6!cxA٫Nd⋳,xJwp$ 0kODr!pvZL=N{^xKM醍@`̗aw\! Ҹ^,]({>(GlfCpIi?i8^%"=A4)؎4rra-u$tNLUaPԿG7soVrs_Sz> 6bZ=~{tOk_)^-]~=v-*2M(1'CۮsQl'B7N)\+UB`L F5gZ -sca3L𡼻$N7m6Z 68~D:syŘ0IaY-|UX.Uvk rVt kc~L\y_8Ԉ47:VE7ix>!0Zaj|.(8b3c&&wOs n 2uQ(ЕdpKAV DznEB@B&Ry# Z641B済2gVt! ".4Ɓh<v *p&d1UM:XTQZMWϼqp#ڲ|P[yD] ϴ< )(zËNʢ= pY݊'B2HSU-Ɔ!̞ o"z2KFYUIe)e. PʂX}ےi::y?cY{jzrgjօj43v*R:hK&|<^j]{fy.1Xi%%@`wo138MA ci$ʲFd\s 1h%#IԨ t>J\-2L2:רw_edw+HH6sL!7bOz˻N-H/6C>ִb u^u’FJIT-oQ 5yw-* wHyh/ 0 Ŝt2^:9J 3L^5:|[ J{Q)߁e1w.ewo[RrJ#U%4sܨn=vԾUXp)5ZO*lN@+48RwK wUj˥"2O-pU/͸ےH<5ٺow5&3(ǖQmE~7]kQ+FVsވtZm.3ʺ0W֘-VMKtR˶*h$+ hw~<Wg#p',F\f_<:*-lwCdn(y.ǣ?Opc%G7VWWWw]/!i-VCpH|KEJ._6l+e2rFv:ұC/:wtG R0\k%F m(XyΗJhD<\'>J=pbIjL_n>Lʨ T&rj̘1{#Xۿ4f[~NZ /-0}$_og@rLc{UIyFp o`! O 6ڰlSؾb[:ÎF =GJ6^dsdL[?g?abfxDni@`ܸMlwew7}9 4+\<#Ll&KmT"WM$wG'yZ=8W.Zy26Wrbtʢݝ7|,3TQD]9icysV莼E=s0;#”әX0 ~yekdu(\ʡa9W&7n.I+؃[ǂ,}RD{2 UbJՅLc"PqrnY7IֿN'rz!&:])?݇!5҆a c"%cB-p~y@7Bf+K8As$|jPuؓ I JP ?> ](V_6ܢ&0q6NX>=8wWD>GQ*l -&ZG2ٙ8Z:UޢV$g|GOH:ު e4Dﵓw[>prgwZFs.0˭7˲22 qQkuY**ϵC_h6$ (=B ;t];SrSm&q#;w>jeX ]`ױ C RF}7*upP1V%)ӱbF@ x$ B)} Djn.Pa;jf׷Z*wu!*0Z5h^^gYRt}37uy_]ej>*ڿwAna&rAw<۾N0?"RN?o7tZZ/+gaq7K5#q~c&kJ9"2,\$!s$SnYh4hNw!eڭȓLևELiy-&Rݺ Gtv;B[zڭ 2(TgH|n&}EoOR>C,) 5³Vcvѯwh2Tn]z~85"`q B“swZƾ BY웏0\@4=:2JX0mvǷ5yC0mv ] +4QQ~ J/Z`i 8N\K㴯BP ĕMN\y!Iڱ2OPv;27?kJmZ.nPG=\~syO.vS.o JLﹼꉣ{; Wq_m.#m!BwZk惎mno7L6r{uPk%ͪ[4`'=ahspX3f 6z[-]$L9m(ff{캰DII{vʒ)O6kѕ܉Qi,:rjYan(;[VgX(B!U3+g5:-F5e·Z7L~%;]G;og`?9`̸v.k/xY| -pҎv_>i6n^ڹ B F_4\0vܰ?rwpÌPo_ˉ]ݢz*Wʿ}Fpo?I̦^b8+WF B}c4YXi v9^EhRƶ9_} QhW*73)(9E*$X]ǂ'bm(*]8uB,~]"7t~֊g?n?4 L4/$c0PA&Z%ERFcv gp@ݦh?+ Ͻ%MNlg^4fuDYtŎf鱛ަߪ6zB-6(E2 ~T_,>`xzY芌aIu5HOp& tWBe:0Og+^4у28+> DGg,ZJ 5']lx,ǝ$^Wܬ`yF:켡,ʶ瓼ƫ˻Y(a.Y'׿<-}IpRWDɸ8)\\1ZR,"km,eu!)iyXa}>-&8otTI\iwϢ e;e)#)^$zl1NmqvNH7lڈg},} *Z6nV Sqݼ#IL>CJL~X`-yeƴ=.wQ2էPb 1+6,M荓OE3(Xexf7:<.rƴx?B}?qC7h(MU^ѐZ$5j e2ŧ揄 <`tR׋ ތX9}c0@7^I! ќ+Otd'Ļ0&ӫW14e%NJ"6S%4ɠ4I H MV3ɏ"yϧHrڮCPL5@fLyv҆,N3ɲb;V'M{w޹j3(!+lwԬ`*B O1|"xy-<iٳb`VxL$l 8J>+}7ñCP^OAwup[k.J~Ъoq'RP+PCSj#?ć#K:Q]HDp݁uz^oGjM!NNZ#x 5!*YMX :Ln~E`L/4J}i^1dg檄.<6U{8Sv|xb8J_69X@(xBSjry[N#kP_h55)Y-CRJo[v/sE ~ilx Fu>WCIۻy۬&|UtV}C p24pcђ)I1Ȋ=mQ܊u4֜x P}8v2NE}& >O|2E~>`TCDܼ B&g~{{,` ux;{UD)9 !fc#f\(C jAԐ\Pod+af=V^AȬNO,o x8FNZ#֠JP}Ѧz7ekȱa:Ƙl,jPxQ@qd4 @!?7>X\V N^Vk8=U(Q+uS4 sBorc m9K+"Dy~-FyAE.^JjjѸ,ZooٲZ}<,v~A|a?y1mBm4zޑ%9)޻ <6m ZB 1Zl aG!'MI/  Z#>":o zʱPRgcёƫэJ"u R*N]Nbvd0D);myUnZKsZCf{W1X$oyLk['̅눏H=tN) p-C@ Z|.id46㸂 l" Ĩ R}4Xf{u,9A5P,v t)so\v.8uBQ5._w:YZ{[9E4ҳ\&Zb5B64ߠ >yYByv漸v,q{<2'9 Ի_fasiV}Z^疕?X?R #S:D˺naM#9x/|,ڸsXT/Joy8iζkOe6hh"a maH B]¨Z4єqA2˺¶͑OCZS஧DF ڋ DJpjЗO4)ݮoߊ?UkJ EmS&KnfȢh lj]8⠎b˫zZr E1$!tEWg\~ЉF]8~|zkTvmT*X([2潱^|ƌDcEM|p*!s_q s3lyN:+H&b6U[XrPYĿ797DlO:rYų)c2dFh2,Ǔ8Q) (Oh3lP똷RBb?qWw(us}NͫvWˡ`e\j*3U܆E'!Kf/?&I#;O-^CqoCY]^ Rq.S?,oAuYsِУ+ypUv/B|/JӠb*YAJ@SAgOʝLڇyq[H%a^[sm K "ԪR(÷#[ȟNwգRpPѴl1̝aw$y-d2Zzq/Z%ZeHtSQbe^IKLiO^SkERNxZoCHls ,KjVdޏG |x{}qD~pb%Ac\dI+d}$'C/y=/.\2/CiX.Q8N/):1Pd];u6g>'! NTĤq,aދrhmh:[<o-|bȨ\ެRE{K Ր⚋Ți$zV:Ys:JIߚLJQ  A-rX J:8Q Щ*n-^bϽf(!y]B>Tح=_o~R+Z0e.롖&-g(@ p@Yhs7W"]2ժ h{}f*wG`S*tτ߽'X/L9ݼD=Xy5!Јxכ~@!AZj޽NPpY!sME, UQ;UR's*^V\^t3he{o%[fM B߯Sf< v[]~}gC#Xҫ4 FЪ>\j:T?XTO>Z'+] 4ځ8dVL1w# m^ﰿuvZBA 1WmZ@@1z0ٞ|a>%UQ2bCs=4Amۇhh2.Ŀ.}qCq.i!&HMT%om*Vũ`y|?Ep_]!рI7fB4MD cgG[ud;徬|d<-..TL#?< }a?߱ ^@~6Fp>[4xqWÒuaz/Aa+-~<񱤱F;cښH_!x9;SGa8sq .6ލ'DM.K/3++*/l~q+jV/Zc`PΒ,@^}.:vLXF@8Z|ZՆ'W(0K;xæiL{ZX yY"N22`g!S .WFg"䆂< f>6,\?NU:T5;7d"I亴y7hw;DD#pꭥ͗Y.q- ?;kl~6mqDʯYW lXiK~{^ǔdNmd:o-\~)Ga9p9&sDjVmW{.(*%kߢCVzM>HtV.8(}+d឴F QGևVke?{ӛ[ rs[KF*Noo~C,k,{MLV0(|7;~Pd18Ds0U +Fh'2,YSd'WwkuF8Y{STM$p!-/_lF3t5}O`P]J250Y/2gxwwj8|XW!;?#U+/QL#p^bdɶڢN%Q;0BI8xJ}H;FB 7(dZ2v&tcTNbRu}UzuKQId I32oUJE`LJ,.(24xQEq %"iL,dcxg1#d͜fmն] 3؁ry93}"xƲHPQ&&2`QlҠ^lROܥ?GX#,4K6 =H)6*"9&hv1Vꪥzj\1.l9̊t5 fH)hsowo65蹆4*=vmmT%`a\m) );̽t={,w-ٺIRC_ӫIOHK [[#kD^+t;wIrT=[c7^RJ,؉0} @y=A|p%39 t©Ģ |&;ØeNԕAai_VY;h.c?rxOxo49L]j$#9xNjOrDj7\:jq#Kt+Iڒ;%gzŜr,#e5di Ԙ!mŨҁRr>/ QS'RDITvi fJ5ٔ#H/l{[6dO]jƽ24f#K3n:R yDyJt-]pХ'$K6Xaтj1Z͘l${:XVYP˵BI?{ԆR@iI5+jwҋ;~^w$P 2M^ }s{'!#Z:ş~Ěًi7#LQu3/ZؿuV|:[nņ0v_iƌ4Ng(37=M``gPF0V})>}5w&DňQ8)7A[Ӷ0WcۋKp.}ù%.xf]Ѵ P)~zJ J<9iIyjь?9a$hzzShӳF!ܪy't#Au[MݏUʞE5H aWh(uВel KHn젮2lDTSV;4g#4b  %+$Be'k]!w6wІthvc?p3"tn/,A#6%]-!GɅG+O%gv23,HB;ҜWF~˕t8inq|]vaU,oPGUbH|́f{(>aJypX>@tg60yJw:ש̓"I#(7<,)/)0YK>kQvJj֣]=A[oZԠm5hW56g$Ѯ]I|Tn$QdG66׭|IHńPu%YT|nڗ~5 bn^9SG2Z;gI4dǕ{G<'5C6J Fe[+T ٍ JS&[6Hs[C]J^$r۔ݩ㟁onPF]5eoʼn|gѴ:)i7CJru:2oюÇx>e|rm>A7ڸ 4鞠;VE^8L%ΎƪLawJ%9WT0wD!uL]%*T~ (K=i3WV:eaݤ[[Qhbݾ2Uz8h%{J|ҕLt5f9]YjLk$XEVpjwH[/b\PS Rx-WDV^x)⠖y?x&hRI[&ZF Z<.K # Ug\(ꍣ,2b% ̆0d_OLlq1V"Db mc}FH&$4`GKFb25 D >̨=vFyրV?1Hk֠;NF21Jy.NxFyR 9F@ށ'FjpBuyv#եcp#iLD<RrE teDižTs;ܻ6: 3+(ц<c$J6y5WP`W~x.,b'qF !B+ٚU( w<: Ss䒽@5Y@U/8TX J5yr5 ZKN}{E״]C&3,K/k* rK=CҨSZ@l>`obGEфU8'Iݭ'ܞ(O\H Iil7+Dm2Od^UNgZdM p*L$kXV'n#3Gb%YmkuV%(zCkEGf?rHij͇4c|5kvP 68yoGER LtH婈,)Ծ`ؘyĮ2bE*ئYDy yzp߳P4Z%褀#N&L~2md"^R=A&|_d "3$4IGJgI)ӯ<F_<h`Z_!L?.Z=ȺuZ[gNy8nLq¢(@bHDI)K<0Ƽ16nV 7ط "+J bnNF;Ɉb.~"LzN6-Ɇ#"xqVo+k+yե|o{/}`dz?]"s}e办iQ척KAz͟D~~绪~LhI…DRRZ. (F +Rt-G&ع$ s{ʥ0@_ C9bxӁ!hvVW+% zey.f,59ܮǢ|q5poy Di{=0mPX3XT`Sxz#K!5(ek;dQ/ [iv_}V$c4, -y+jAt)5H4fZ -R*DhcЊ8(ɂHblݷ?ⁱ^p?9**W(M'-٭9[7Hkϼe=+ڼ|>K_xJ53SUO$6y/Jzv˦h{?@.7~s}=5cPe_\_ ~! [zZx2_9z3>.|'3}h?A~{dp:8~_\p_r|6W_޾tW/WéSË n_?t~=|9.<7T^W_^ <}̦W8w/Cܓ,O˳帴;Sǯ[L'}1ӳOLG'30KP/MRo.)wpT[dvYvPsoD_L ,y9;4&{g5K-W-ٿ I5NU_ILopוֺ~MZƲ..-'(p \V]ZUz@tXMi#i@~vwj%F󡤐 q.iuQ~韐 gF+$M4((Jy2C(mط6Nr3R|hlk}(ZB1Jצ&H86 :7:㋺n4dëBmc^M۟UtfyU9SoNnl.YlT-3~fm=h,p+X_]gx+v~xMd{/FW{pW ftw6 ]rf{ f Xj7B#|Q&m)=uɄ=t_w}NWzFcPM?U%tMysuHzʒ}m<жmhWtJoo[7}֭)BTu!BnX[N@}Đf1/m\Xsޗ/płkĠ$-UZ⫙Q@?LctWx@U$s-w|^OGxtuٛL/-oG[u-nCkrAl+g 5LYj 3\4_lti~m#PU~Fb/@ϑllt gՕ.:-^܄Z5)ݷOPP/('’H}Tx @=(-yΪy!0$Sm&1Icefц1f.Y)<>nfb\#'ɝ\ -r;MFJ#dl8Jlp$pyV8D_@7/؄dJ4T"TY(Bct*JlWVZXl<H`@0i6RJ bʖ,o澿H6RWꭍmnn>!uDH}pw1D!V95+sxq)+^ޮ{y.zdV- K[,]jd wU]%[]ud V-]'[$0"u*W Jᚕµ,]pcXHVtvU \5\ۍ5s<6qlq XbMB8i"Z)`MmM~z{6ٹR@e %T \o*qj<xv㻳{kݝx%Ii,Onk-hDEƆmaIa|R9RЪynu!`5_d7!I4] Q<`wTt{-K"a^V!PD{=v#3_LY%PBI:#P%W1t'ցIɳ>l% ʴ-X`*ũI(B k$q)+C 0IH`걀#8˚Icv&(-Sa[s# A:ٝ-_;2 3etshSRSCpZO`K`j0Gi+xOc"!^Ƴk49; Řw4'{K`X2[@ T-Dp` BLr0Z8*#s*.RY AZݝ†@P mnb2vWy' rF|rOI>=-QjոV`J\gݸIZDFaROEln Ƙ[aߠ#L|IQN.2ʇ! H)6E7N F}iEsZg֛rd{ G#"L]6"{>+mcXjr+ aA\ ~- S{S) es-8t"{C*˷!z/\?,[C(ivX)#ziѶ ;I{yC0 G)׻u p18:T Hj9`?!u3-3™TEFCzVM~PPBKlj>n-ՁZCfPIGڤ!Z4*% KM[,S,\bH j awZQ9ލjF"@I Fe$OMy T ArC)9YbW2j,"($E2"TkOE7*k)@RK) ҼKA;-}Zh2굴 pH-e4LK=旝>f-*LKysQwQ}J\vZ"]r~Z_߳]}V(jEŮ$ip$oy^xd|㏭0^W@P-)kWA&PI|Ov{|n9wbUjґKʂ9*ީ/̀+(#'sY`Lb_j~tRRBHtֈ哥D!QDv@ 3҄pN*KTu:s ,A*BVwC޴e::HkԱs1r,'6B(&f 0%7ۥCc?*`N2b`J'd! )\Eą_EKBb 0)n! ֫y lw21!1D(k2'x"& iՂ_I( "BX" JKS"D VGbdE|ou!var/home/core/zuul-output/logs/kubelet.log0000644000000000000000005114563015150353622017704 0ustar rootrootFeb 27 16:23:06 crc systemd[1]: Starting Kubernetes Kubelet... Feb 27 16:23:06 crc restorecon[4697]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:06 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 16:23:07 crc restorecon[4697]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 16:23:07 crc restorecon[4697]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 27 16:23:08 crc kubenswrapper[4814]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 27 16:23:08 crc kubenswrapper[4814]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 27 16:23:08 crc kubenswrapper[4814]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 27 16:23:08 crc kubenswrapper[4814]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 27 16:23:08 crc kubenswrapper[4814]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 27 16:23:08 crc kubenswrapper[4814]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.074326 4814 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100625 4814 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100686 4814 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100698 4814 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100710 4814 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100720 4814 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100729 4814 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100738 4814 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100746 4814 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100754 4814 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100762 4814 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100771 4814 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100780 4814 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100789 4814 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100797 4814 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100805 4814 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100813 4814 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100821 4814 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100829 4814 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100836 4814 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100844 4814 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100852 4814 feature_gate.go:330] unrecognized feature gate: Example Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100860 4814 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100867 4814 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100876 4814 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100883 4814 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100891 4814 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100903 4814 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100915 4814 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100924 4814 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100933 4814 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100942 4814 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100950 4814 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100958 4814 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100966 4814 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.100974 4814 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101013 4814 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101024 4814 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101033 4814 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101042 4814 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101050 4814 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101058 4814 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101066 4814 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101075 4814 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101083 4814 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101091 4814 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101103 4814 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101114 4814 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101123 4814 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101131 4814 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101139 4814 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101146 4814 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101155 4814 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101163 4814 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101171 4814 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101178 4814 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101186 4814 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101195 4814 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101203 4814 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101211 4814 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101218 4814 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101229 4814 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101238 4814 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101248 4814 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101297 4814 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101306 4814 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101315 4814 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101324 4814 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101333 4814 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101342 4814 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101350 4814 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.101358 4814 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101523 4814 flags.go:64] FLAG: --address="0.0.0.0" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101541 4814 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101556 4814 flags.go:64] FLAG: --anonymous-auth="true" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101568 4814 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101580 4814 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101589 4814 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101604 4814 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101615 4814 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101624 4814 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101634 4814 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101643 4814 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101654 4814 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101664 4814 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101673 4814 flags.go:64] FLAG: --cgroup-root="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101682 4814 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101691 4814 flags.go:64] FLAG: --client-ca-file="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101699 4814 flags.go:64] FLAG: --cloud-config="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101709 4814 flags.go:64] FLAG: --cloud-provider="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101717 4814 flags.go:64] FLAG: --cluster-dns="[]" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101728 4814 flags.go:64] FLAG: --cluster-domain="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101736 4814 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101746 4814 flags.go:64] FLAG: --config-dir="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101755 4814 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101765 4814 flags.go:64] FLAG: --container-log-max-files="5" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101776 4814 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101785 4814 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101794 4814 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101803 4814 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101813 4814 flags.go:64] FLAG: --contention-profiling="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101822 4814 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101832 4814 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101842 4814 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101850 4814 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101863 4814 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101872 4814 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101881 4814 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101890 4814 flags.go:64] FLAG: --enable-load-reader="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101899 4814 flags.go:64] FLAG: --enable-server="true" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101912 4814 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101925 4814 flags.go:64] FLAG: --event-burst="100" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101935 4814 flags.go:64] FLAG: --event-qps="50" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101944 4814 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101953 4814 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101962 4814 flags.go:64] FLAG: --eviction-hard="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101974 4814 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101983 4814 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.101992 4814 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102002 4814 flags.go:64] FLAG: --eviction-soft="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102012 4814 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102020 4814 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102030 4814 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102038 4814 flags.go:64] FLAG: --experimental-mounter-path="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102047 4814 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102056 4814 flags.go:64] FLAG: --fail-swap-on="true" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102065 4814 flags.go:64] FLAG: --feature-gates="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102085 4814 flags.go:64] FLAG: --file-check-frequency="20s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102094 4814 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102104 4814 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102113 4814 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102123 4814 flags.go:64] FLAG: --healthz-port="10248" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102132 4814 flags.go:64] FLAG: --help="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102141 4814 flags.go:64] FLAG: --hostname-override="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102150 4814 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102159 4814 flags.go:64] FLAG: --http-check-frequency="20s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102169 4814 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102178 4814 flags.go:64] FLAG: --image-credential-provider-config="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102187 4814 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102196 4814 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102204 4814 flags.go:64] FLAG: --image-service-endpoint="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102213 4814 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102224 4814 flags.go:64] FLAG: --kube-api-burst="100" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102233 4814 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102243 4814 flags.go:64] FLAG: --kube-api-qps="50" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102279 4814 flags.go:64] FLAG: --kube-reserved="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102290 4814 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102298 4814 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102308 4814 flags.go:64] FLAG: --kubelet-cgroups="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102316 4814 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102325 4814 flags.go:64] FLAG: --lock-file="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102334 4814 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102343 4814 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102352 4814 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102366 4814 flags.go:64] FLAG: --log-json-split-stream="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102376 4814 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102385 4814 flags.go:64] FLAG: --log-text-split-stream="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102394 4814 flags.go:64] FLAG: --logging-format="text" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102403 4814 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102412 4814 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102421 4814 flags.go:64] FLAG: --manifest-url="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102430 4814 flags.go:64] FLAG: --manifest-url-header="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102443 4814 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102452 4814 flags.go:64] FLAG: --max-open-files="1000000" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102463 4814 flags.go:64] FLAG: --max-pods="110" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102472 4814 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102481 4814 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102490 4814 flags.go:64] FLAG: --memory-manager-policy="None" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102499 4814 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102508 4814 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102517 4814 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102526 4814 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102546 4814 flags.go:64] FLAG: --node-status-max-images="50" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102556 4814 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102565 4814 flags.go:64] FLAG: --oom-score-adj="-999" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102575 4814 flags.go:64] FLAG: --pod-cidr="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102583 4814 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102597 4814 flags.go:64] FLAG: --pod-manifest-path="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102637 4814 flags.go:64] FLAG: --pod-max-pids="-1" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102648 4814 flags.go:64] FLAG: --pods-per-core="0" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102657 4814 flags.go:64] FLAG: --port="10250" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102667 4814 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102676 4814 flags.go:64] FLAG: --provider-id="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102684 4814 flags.go:64] FLAG: --qos-reserved="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102693 4814 flags.go:64] FLAG: --read-only-port="10255" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102702 4814 flags.go:64] FLAG: --register-node="true" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102711 4814 flags.go:64] FLAG: --register-schedulable="true" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102721 4814 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102735 4814 flags.go:64] FLAG: --registry-burst="10" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102744 4814 flags.go:64] FLAG: --registry-qps="5" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102753 4814 flags.go:64] FLAG: --reserved-cpus="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102764 4814 flags.go:64] FLAG: --reserved-memory="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102775 4814 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102785 4814 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102794 4814 flags.go:64] FLAG: --rotate-certificates="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102803 4814 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102812 4814 flags.go:64] FLAG: --runonce="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102820 4814 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102830 4814 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102840 4814 flags.go:64] FLAG: --seccomp-default="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102849 4814 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102858 4814 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102867 4814 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102877 4814 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102886 4814 flags.go:64] FLAG: --storage-driver-password="root" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102895 4814 flags.go:64] FLAG: --storage-driver-secure="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102904 4814 flags.go:64] FLAG: --storage-driver-table="stats" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102914 4814 flags.go:64] FLAG: --storage-driver-user="root" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102923 4814 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102932 4814 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102941 4814 flags.go:64] FLAG: --system-cgroups="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102950 4814 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102964 4814 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102973 4814 flags.go:64] FLAG: --tls-cert-file="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.102982 4814 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.103453 4814 flags.go:64] FLAG: --tls-min-version="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.103464 4814 flags.go:64] FLAG: --tls-private-key-file="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.103473 4814 flags.go:64] FLAG: --topology-manager-policy="none" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.103483 4814 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.103492 4814 flags.go:64] FLAG: --topology-manager-scope="container" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.103502 4814 flags.go:64] FLAG: --v="2" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.103514 4814 flags.go:64] FLAG: --version="false" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.103525 4814 flags.go:64] FLAG: --vmodule="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.103536 4814 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.103547 4814 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103750 4814 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103760 4814 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103770 4814 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103779 4814 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103788 4814 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103796 4814 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103804 4814 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103814 4814 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103824 4814 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103832 4814 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103841 4814 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103850 4814 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103858 4814 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103867 4814 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103876 4814 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103884 4814 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103893 4814 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103901 4814 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103909 4814 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103917 4814 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103925 4814 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103933 4814 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103941 4814 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103949 4814 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103956 4814 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103965 4814 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103972 4814 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103980 4814 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103988 4814 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.103996 4814 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104004 4814 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104012 4814 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104019 4814 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104028 4814 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104036 4814 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104043 4814 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104051 4814 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104062 4814 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104074 4814 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104083 4814 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104093 4814 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104102 4814 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104111 4814 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104134 4814 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104145 4814 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104154 4814 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104171 4814 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104179 4814 feature_gate.go:330] unrecognized feature gate: Example Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104188 4814 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104197 4814 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104207 4814 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104215 4814 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104223 4814 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104231 4814 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104239 4814 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104279 4814 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104290 4814 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104299 4814 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104307 4814 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104315 4814 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104323 4814 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104331 4814 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104340 4814 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104348 4814 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104356 4814 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104364 4814 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104371 4814 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104380 4814 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104387 4814 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104395 4814 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.104403 4814 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.104426 4814 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.117347 4814 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.117399 4814 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117526 4814 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117539 4814 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117548 4814 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117557 4814 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117565 4814 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117573 4814 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117582 4814 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117590 4814 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117598 4814 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117605 4814 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117613 4814 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117621 4814 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117632 4814 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117643 4814 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117652 4814 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117662 4814 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117672 4814 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117682 4814 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117690 4814 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117699 4814 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117706 4814 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117715 4814 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117723 4814 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117730 4814 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117738 4814 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117746 4814 feature_gate.go:330] unrecognized feature gate: Example Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117755 4814 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117763 4814 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117771 4814 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117778 4814 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117788 4814 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117799 4814 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117809 4814 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117818 4814 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117828 4814 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117836 4814 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117844 4814 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117852 4814 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117862 4814 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117872 4814 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117882 4814 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117890 4814 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117899 4814 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117908 4814 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117916 4814 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117924 4814 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117932 4814 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117940 4814 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117949 4814 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117957 4814 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117965 4814 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117972 4814 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117980 4814 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117988 4814 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.117996 4814 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118004 4814 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118011 4814 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118019 4814 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118026 4814 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118037 4814 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118046 4814 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118054 4814 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118062 4814 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118069 4814 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118077 4814 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118085 4814 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118095 4814 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118105 4814 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118114 4814 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118124 4814 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118135 4814 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.118148 4814 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118393 4814 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118404 4814 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118413 4814 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118427 4814 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118439 4814 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118448 4814 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118457 4814 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118467 4814 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118476 4814 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118484 4814 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118492 4814 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118501 4814 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118509 4814 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118517 4814 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118525 4814 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118533 4814 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118540 4814 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118548 4814 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118556 4814 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118564 4814 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118571 4814 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118579 4814 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118586 4814 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118595 4814 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118619 4814 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118628 4814 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118636 4814 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118644 4814 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118652 4814 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118659 4814 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118667 4814 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118675 4814 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118682 4814 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118690 4814 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118699 4814 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118707 4814 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118715 4814 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118723 4814 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118731 4814 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118740 4814 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118749 4814 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118757 4814 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118764 4814 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118773 4814 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118781 4814 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118788 4814 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118796 4814 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118804 4814 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118812 4814 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118825 4814 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118838 4814 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118851 4814 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118861 4814 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118871 4814 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118881 4814 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118891 4814 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118900 4814 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118908 4814 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118916 4814 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118924 4814 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118932 4814 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118940 4814 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118950 4814 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118960 4814 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118969 4814 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118977 4814 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118985 4814 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.118993 4814 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.119001 4814 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.119009 4814 feature_gate.go:330] unrecognized feature gate: Example Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.119018 4814 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.119031 4814 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.119317 4814 server.go:940] "Client rotation is on, will bootstrap in background" Feb 27 16:23:08 crc kubenswrapper[4814]: E0227 16:23:08.125849 4814 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.138337 4814 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.138491 4814 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.141324 4814 server.go:997] "Starting client certificate rotation" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.141370 4814 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.142284 4814 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.210828 4814 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 27 16:23:08 crc kubenswrapper[4814]: E0227 16:23:08.214056 4814 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.215270 4814 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.270225 4814 log.go:25] "Validated CRI v1 runtime API" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.329290 4814 log.go:25] "Validated CRI v1 image API" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.332937 4814 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.341842 4814 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-27-16-18-12-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.341877 4814 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.361951 4814 manager.go:217] Machine: {Timestamp:2026-02-27 16:23:08.357629231 +0000 UTC m=+0.810254101 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:1f184c22-f8d8-46af-88b2-5c278287a2f3 BootID:9eed0068-32e1-4174-93ec-451c02789f32 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:89:bf:49 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:89:bf:49 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:9b:3f:8e Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:9f:a4:a6 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:62:e4:fc Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:2d:a6:73 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:12:e6:6b:a0:50:34 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:8e:14:54:f4:45:34 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.362227 4814 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.362482 4814 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.366751 4814 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.367030 4814 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.367071 4814 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.367386 4814 topology_manager.go:138] "Creating topology manager with none policy" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.367404 4814 container_manager_linux.go:303] "Creating device plugin manager" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.368101 4814 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.368148 4814 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.368397 4814 state_mem.go:36] "Initialized new in-memory state store" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.368540 4814 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.371981 4814 kubelet.go:418] "Attempting to sync node with API server" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.372015 4814 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.372043 4814 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.372059 4814 kubelet.go:324] "Adding apiserver pod source" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.372073 4814 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.376241 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Feb 27 16:23:08 crc kubenswrapper[4814]: E0227 16:23:08.376399 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.376452 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Feb 27 16:23:08 crc kubenswrapper[4814]: E0227 16:23:08.376559 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.377878 4814 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.379635 4814 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.381334 4814 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.383117 4814 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.383147 4814 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.383157 4814 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.383167 4814 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.383181 4814 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.383189 4814 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.383196 4814 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.383208 4814 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.383216 4814 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.383223 4814 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.383275 4814 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.383286 4814 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.385655 4814 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.386142 4814 server.go:1280] "Started kubelet" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.396478 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.396143 4814 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.396621 4814 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.397470 4814 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 27 16:23:08 crc systemd[1]: Started Kubernetes Kubelet. Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.426412 4814 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.426482 4814 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.426631 4814 server.go:460] "Adding debug handlers to kubelet server" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.427046 4814 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.427121 4814 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 27 16:23:08 crc kubenswrapper[4814]: E0227 16:23:08.427057 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.427229 4814 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 27 16:23:08 crc kubenswrapper[4814]: E0227 16:23:08.427781 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" interval="200ms" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.428289 4814 factory.go:55] Registering systemd factory Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.428312 4814 factory.go:221] Registration of the systemd container factory successfully Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.428403 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Feb 27 16:23:08 crc kubenswrapper[4814]: E0227 16:23:08.428503 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.428620 4814 factory.go:153] Registering CRI-O factory Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.428650 4814 factory.go:221] Registration of the crio container factory successfully Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.428806 4814 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.428882 4814 factory.go:103] Registering Raw factory Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.428941 4814 manager.go:1196] Started watching for new ooms in manager Feb 27 16:23:08 crc kubenswrapper[4814]: E0227 16:23:08.427199 4814 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.189:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1898270c965ddbb2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.386114482 +0000 UTC m=+0.838739312,LastTimestamp:2026-02-27 16:23:08.386114482 +0000 UTC m=+0.838739312,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.430002 4814 manager.go:319] Starting recovery of all containers Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434546 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434610 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434625 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434636 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434647 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434658 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434670 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434682 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434695 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434718 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434754 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434764 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434774 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434786 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434797 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434811 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434821 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434830 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434844 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434854 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434864 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434877 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434888 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434898 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434908 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.434918 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435075 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435089 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435100 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435114 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435124 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435163 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435175 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435186 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435220 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435233 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435242 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435269 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435279 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435288 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435321 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435330 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.435340 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438489 4814 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438526 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438604 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438621 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438635 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438648 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438662 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438676 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438688 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438703 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438717 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438729 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438741 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438754 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438766 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438777 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438790 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438802 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438813 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438824 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438834 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438845 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438856 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438865 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438876 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438887 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438897 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438909 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438921 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438931 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438941 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.438953 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440090 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440105 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440116 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440128 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440138 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440148 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440161 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440176 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440189 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440200 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440210 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440222 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440232 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440272 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440288 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440301 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440313 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440344 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440357 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440375 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440387 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440404 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440417 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440428 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440441 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440454 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440466 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440480 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440494 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440512 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440547 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440563 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440575 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440592 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440606 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440618 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440630 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440642 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440654 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440668 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440687 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440700 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440713 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440724 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440736 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440755 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440766 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440779 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440791 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440810 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440821 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440834 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440846 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440858 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440877 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440888 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440907 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440918 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440930 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440944 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440964 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440976 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440987 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.440998 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441015 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441027 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441042 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441053 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441067 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441078 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441108 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441124 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441153 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441165 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441178 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441190 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441201 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441213 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441223 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441234 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441264 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441277 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441289 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441302 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441319 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441330 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441344 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441356 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441367 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441382 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441393 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441888 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441934 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.441966 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.442020 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.442054 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.442097 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.442114 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.442136 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.442171 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.442185 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.442207 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.442221 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.442426 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443099 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443204 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443243 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443296 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443318 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443356 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443377 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443420 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443441 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443480 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443509 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443596 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443637 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443669 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443691 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443719 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443748 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443776 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443797 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443826 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443854 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443885 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443912 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443948 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.443970 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.444037 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.444066 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.444091 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.444119 4814 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.444137 4814 reconstruct.go:97] "Volume reconstruction finished" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.444149 4814 reconciler.go:26] "Reconciler: start to sync state" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.455965 4814 manager.go:324] Recovery completed Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.465900 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.467460 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.467750 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.467762 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.471372 4814 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.471395 4814 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.471417 4814 state_mem.go:36] "Initialized new in-memory state store" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.484492 4814 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.486249 4814 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.486321 4814 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.486357 4814 kubelet.go:2335] "Starting kubelet main sync loop" Feb 27 16:23:08 crc kubenswrapper[4814]: E0227 16:23:08.486464 4814 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 27 16:23:08 crc kubenswrapper[4814]: W0227 16:23:08.491094 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Feb 27 16:23:08 crc kubenswrapper[4814]: E0227 16:23:08.491181 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.496265 4814 policy_none.go:49] "None policy: Start" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.497012 4814 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.497044 4814 state_mem.go:35] "Initializing new in-memory state store" Feb 27 16:23:08 crc kubenswrapper[4814]: E0227 16:23:08.527669 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.564797 4814 manager.go:334] "Starting Device Plugin manager" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.564874 4814 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.564894 4814 server.go:79] "Starting device plugin registration server" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.565581 4814 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.565638 4814 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.565872 4814 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.565996 4814 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.566016 4814 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 27 16:23:08 crc kubenswrapper[4814]: E0227 16:23:08.574958 4814 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.587387 4814 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.587469 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.589010 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.589038 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.589048 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.589190 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.590106 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.590173 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.590192 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.593997 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.594047 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.594114 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.594313 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.594349 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.595244 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.595312 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.595324 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.595391 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.595417 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.595433 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.596294 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.596320 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.596331 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.596557 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.596761 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.596843 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.597643 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.597762 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.597881 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.598137 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.598234 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.598314 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.599730 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.599821 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.599891 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.599938 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.600046 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.600068 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.600312 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.600403 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.600430 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.600683 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.600720 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.601745 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.601776 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.601820 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:08 crc kubenswrapper[4814]: E0227 16:23:08.628686 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" interval="400ms" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.645770 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.645845 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.645929 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.646056 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.646130 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.646179 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.646278 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.646337 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.646366 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.646393 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.646472 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.646503 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.646644 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.646673 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.646732 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.665753 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.669281 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.669327 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.669342 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.669372 4814 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 16:23:08 crc kubenswrapper[4814]: E0227 16:23:08.669881 4814 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.189:6443: connect: connection refused" node="crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.747957 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.748023 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.748059 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.748088 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.748121 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.748149 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.748177 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.748205 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.748238 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.748328 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.748371 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.748400 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.748427 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.748457 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.748485 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.748935 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.749023 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.749049 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.749076 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.749033 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.749104 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.749134 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.749150 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.749160 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.749178 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.749185 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.749210 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.749213 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.749241 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.749244 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.870855 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.872293 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.872358 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.872370 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.872400 4814 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 16:23:08 crc kubenswrapper[4814]: E0227 16:23:08.872985 4814 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.189:6443: connect: connection refused" node="crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.941035 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.951334 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.978445 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:08 crc kubenswrapper[4814]: I0227 16:23:08.985950 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:09 crc kubenswrapper[4814]: I0227 16:23:09.020967 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 16:23:09 crc kubenswrapper[4814]: W0227 16:23:09.027922 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-e791943bb8fbce3462866dd8862f9e97d30cf6d524b6192a648e363f902472cd WatchSource:0}: Error finding container e791943bb8fbce3462866dd8862f9e97d30cf6d524b6192a648e363f902472cd: Status 404 returned error can't find the container with id e791943bb8fbce3462866dd8862f9e97d30cf6d524b6192a648e363f902472cd Feb 27 16:23:09 crc kubenswrapper[4814]: E0227 16:23:09.029908 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" interval="800ms" Feb 27 16:23:09 crc kubenswrapper[4814]: W0227 16:23:09.032863 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-b40aa30f57f677fc0f71d53c56281a521c28769be334653b9ce2670368876ff3 WatchSource:0}: Error finding container b40aa30f57f677fc0f71d53c56281a521c28769be334653b9ce2670368876ff3: Status 404 returned error can't find the container with id b40aa30f57f677fc0f71d53c56281a521c28769be334653b9ce2670368876ff3 Feb 27 16:23:09 crc kubenswrapper[4814]: W0227 16:23:09.034699 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-36e3eb7b1c19b193a10a1389c2c269742f438d2379b3f210f70d37e389a0ef76 WatchSource:0}: Error finding container 36e3eb7b1c19b193a10a1389c2c269742f438d2379b3f210f70d37e389a0ef76: Status 404 returned error can't find the container with id 36e3eb7b1c19b193a10a1389c2c269742f438d2379b3f210f70d37e389a0ef76 Feb 27 16:23:09 crc kubenswrapper[4814]: W0227 16:23:09.045884 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-658056ca9ad67caa8aa9e02aa7097002cc9f05e816564f2eca6165f7c503f536 WatchSource:0}: Error finding container 658056ca9ad67caa8aa9e02aa7097002cc9f05e816564f2eca6165f7c503f536: Status 404 returned error can't find the container with id 658056ca9ad67caa8aa9e02aa7097002cc9f05e816564f2eca6165f7c503f536 Feb 27 16:23:09 crc kubenswrapper[4814]: W0227 16:23:09.046772 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-16e8c9b49c9ac95307bf1e87b64e21464f7874b40e15f9d72f87902981c4ca89 WatchSource:0}: Error finding container 16e8c9b49c9ac95307bf1e87b64e21464f7874b40e15f9d72f87902981c4ca89: Status 404 returned error can't find the container with id 16e8c9b49c9ac95307bf1e87b64e21464f7874b40e15f9d72f87902981c4ca89 Feb 27 16:23:09 crc kubenswrapper[4814]: I0227 16:23:09.274000 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:09 crc kubenswrapper[4814]: I0227 16:23:09.276009 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:09 crc kubenswrapper[4814]: I0227 16:23:09.276080 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:09 crc kubenswrapper[4814]: I0227 16:23:09.276104 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:09 crc kubenswrapper[4814]: I0227 16:23:09.276147 4814 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 16:23:09 crc kubenswrapper[4814]: E0227 16:23:09.276721 4814 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.189:6443: connect: connection refused" node="crc" Feb 27 16:23:09 crc kubenswrapper[4814]: I0227 16:23:09.397804 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Feb 27 16:23:09 crc kubenswrapper[4814]: I0227 16:23:09.490967 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"658056ca9ad67caa8aa9e02aa7097002cc9f05e816564f2eca6165f7c503f536"} Feb 27 16:23:09 crc kubenswrapper[4814]: I0227 16:23:09.492231 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b40aa30f57f677fc0f71d53c56281a521c28769be334653b9ce2670368876ff3"} Feb 27 16:23:09 crc kubenswrapper[4814]: I0227 16:23:09.493184 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"36e3eb7b1c19b193a10a1389c2c269742f438d2379b3f210f70d37e389a0ef76"} Feb 27 16:23:09 crc kubenswrapper[4814]: I0227 16:23:09.494233 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e791943bb8fbce3462866dd8862f9e97d30cf6d524b6192a648e363f902472cd"} Feb 27 16:23:09 crc kubenswrapper[4814]: I0227 16:23:09.495538 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"16e8c9b49c9ac95307bf1e87b64e21464f7874b40e15f9d72f87902981c4ca89"} Feb 27 16:23:09 crc kubenswrapper[4814]: W0227 16:23:09.535826 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Feb 27 16:23:09 crc kubenswrapper[4814]: E0227 16:23:09.535928 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Feb 27 16:23:09 crc kubenswrapper[4814]: W0227 16:23:09.738926 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Feb 27 16:23:09 crc kubenswrapper[4814]: E0227 16:23:09.741263 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Feb 27 16:23:09 crc kubenswrapper[4814]: E0227 16:23:09.831994 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" interval="1.6s" Feb 27 16:23:09 crc kubenswrapper[4814]: W0227 16:23:09.924007 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Feb 27 16:23:09 crc kubenswrapper[4814]: E0227 16:23:09.924118 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Feb 27 16:23:09 crc kubenswrapper[4814]: W0227 16:23:09.942794 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Feb 27 16:23:09 crc kubenswrapper[4814]: E0227 16:23:09.942947 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.076910 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.080080 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.080337 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.080368 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.080414 4814 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 16:23:10 crc kubenswrapper[4814]: E0227 16:23:10.081149 4814 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.189:6443: connect: connection refused" node="crc" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.364570 4814 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 16:23:10 crc kubenswrapper[4814]: E0227 16:23:10.365762 4814 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.397854 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.502243 4814 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b" exitCode=0 Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.502446 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b"} Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.502493 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.503836 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.503894 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.503916 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.506992 4814 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82" exitCode=0 Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.507098 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82"} Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.507232 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.508958 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.509011 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.509033 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.512523 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b86c2da64ed7a7f6af14d9fdd54dbb1ace65f683dc22f739afbffa9f1cd6d52f"} Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.512573 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9fbf88717d3b766e6b99c0ec623769bd67642fe95222e221d59f15e0e2fdf3bd"} Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.512590 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b5ba138e1d2cbeab8ba50d4d02b246c31cb1226615131fa4b074a08d615ca4f8"} Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.512601 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"501ea405666a49080dd2af178db81a992757c23da06fedd1a6036047880f3f7a"} Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.512612 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.513876 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.513904 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.513924 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.514384 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.514850 4814 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="f14102163d15d2563d01aa8d820f93aa6c36a244adb644431399aa94fea3b568" exitCode=0 Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.514951 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.514970 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"f14102163d15d2563d01aa8d820f93aa6c36a244adb644431399aa94fea3b568"} Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.515477 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.515519 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.515538 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.516357 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.516411 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.516431 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.516553 4814 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1" exitCode=0 Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.516592 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1"} Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.516723 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.517606 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.517641 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:10 crc kubenswrapper[4814]: I0227 16:23:10.517664 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.399229 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Feb 27 16:23:11 crc kubenswrapper[4814]: E0227 16:23:11.433326 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" interval="3.2s" Feb 27 16:23:11 crc kubenswrapper[4814]: W0227 16:23:11.461278 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Feb 27 16:23:11 crc kubenswrapper[4814]: E0227 16:23:11.461390 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.523460 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.523840 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e7d1c14a07903094a68c4b62d08ede23259920fb342412bc54a5d1ce0ea92b25"} Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.524573 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.524612 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.524623 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.526718 4814 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915" exitCode=0 Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.526838 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915"} Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.526844 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.527783 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.527841 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.527852 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.529963 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8bc29a3f6d746e0eda7be7e511df92aa978993ac600a788820e292603b328434"} Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.529984 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.529993 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"34064bf25c31b4fb36554561a8cf9109191e950be41e22c71e9d4773e0e66ab4"} Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.530006 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fe23574a71dd13340a8be7b9dae77410b55fd1daa11a58eb8ed914fb44cea829"} Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.530747 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.530776 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.530788 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.533236 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced"} Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.533324 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5"} Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.533343 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479"} Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.533351 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.534750 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.534780 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.534818 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.682239 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.683627 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.683656 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.683664 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:11 crc kubenswrapper[4814]: I0227 16:23:11.683686 4814 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 16:23:11 crc kubenswrapper[4814]: E0227 16:23:11.684193 4814 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.189:6443: connect: connection refused" node="crc" Feb 27 16:23:11 crc kubenswrapper[4814]: W0227 16:23:11.807046 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Feb 27 16:23:11 crc kubenswrapper[4814]: E0227 16:23:11.807175 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Feb 27 16:23:12 crc kubenswrapper[4814]: W0227 16:23:12.076082 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Feb 27 16:23:12 crc kubenswrapper[4814]: E0227 16:23:12.076608 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.539709 4814 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224" exitCode=0 Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.539808 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224"} Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.539964 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.541383 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.541429 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.541447 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.544833 4814 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.544894 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.545161 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"999cf79255061bae18e7033b5e5cbbe73b0f3b85421b580972e97ed99db6b363"} Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.545199 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51"} Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.545201 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.545539 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.546316 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.546337 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.546370 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.546347 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.546496 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.546389 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.547606 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.547641 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:12 crc kubenswrapper[4814]: I0227 16:23:12.547656 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:13 crc kubenswrapper[4814]: I0227 16:23:13.200304 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:13 crc kubenswrapper[4814]: I0227 16:23:13.466238 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:13 crc kubenswrapper[4814]: I0227 16:23:13.552180 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87"} Feb 27 16:23:13 crc kubenswrapper[4814]: I0227 16:23:13.552238 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3"} Feb 27 16:23:13 crc kubenswrapper[4814]: I0227 16:23:13.552281 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5"} Feb 27 16:23:13 crc kubenswrapper[4814]: I0227 16:23:13.552297 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733"} Feb 27 16:23:13 crc kubenswrapper[4814]: I0227 16:23:13.552470 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:13 crc kubenswrapper[4814]: I0227 16:23:13.552578 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:13 crc kubenswrapper[4814]: I0227 16:23:13.553442 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:13 crc kubenswrapper[4814]: I0227 16:23:13.553481 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:13 crc kubenswrapper[4814]: I0227 16:23:13.553494 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.386631 4814 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.500122 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.500486 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.502281 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.502344 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.502360 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.560728 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.561609 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.561914 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b"} Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.562706 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.562781 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.562808 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.562923 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.562967 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.562983 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.885338 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.887476 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.887608 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.887692 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:14 crc kubenswrapper[4814]: I0227 16:23:14.887821 4814 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 16:23:15 crc kubenswrapper[4814]: I0227 16:23:15.563400 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:15 crc kubenswrapper[4814]: I0227 16:23:15.564543 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:15 crc kubenswrapper[4814]: I0227 16:23:15.564630 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:15 crc kubenswrapper[4814]: I0227 16:23:15.564651 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:16 crc kubenswrapper[4814]: I0227 16:23:16.388203 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:16 crc kubenswrapper[4814]: I0227 16:23:16.388443 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:16 crc kubenswrapper[4814]: I0227 16:23:16.390024 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:16 crc kubenswrapper[4814]: I0227 16:23:16.390110 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:16 crc kubenswrapper[4814]: I0227 16:23:16.390130 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:16 crc kubenswrapper[4814]: I0227 16:23:16.554603 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:16 crc kubenswrapper[4814]: I0227 16:23:16.565061 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:16 crc kubenswrapper[4814]: I0227 16:23:16.566224 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:16 crc kubenswrapper[4814]: I0227 16:23:16.568773 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:16 crc kubenswrapper[4814]: I0227 16:23:16.568837 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:16 crc kubenswrapper[4814]: I0227 16:23:16.568857 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:17 crc kubenswrapper[4814]: I0227 16:23:17.568822 4814 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 16:23:17 crc kubenswrapper[4814]: I0227 16:23:17.568891 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:17 crc kubenswrapper[4814]: I0227 16:23:17.570106 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:17 crc kubenswrapper[4814]: I0227 16:23:17.570135 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:17 crc kubenswrapper[4814]: I0227 16:23:17.570145 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:17 crc kubenswrapper[4814]: I0227 16:23:17.688930 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:18 crc kubenswrapper[4814]: I0227 16:23:18.571150 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:18 crc kubenswrapper[4814]: I0227 16:23:18.572471 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:18 crc kubenswrapper[4814]: I0227 16:23:18.572509 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:18 crc kubenswrapper[4814]: I0227 16:23:18.572523 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:18 crc kubenswrapper[4814]: E0227 16:23:18.575050 4814 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 16:23:19 crc kubenswrapper[4814]: I0227 16:23:19.544088 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 27 16:23:19 crc kubenswrapper[4814]: I0227 16:23:19.544466 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:19 crc kubenswrapper[4814]: I0227 16:23:19.546044 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:19 crc kubenswrapper[4814]: I0227 16:23:19.546112 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:19 crc kubenswrapper[4814]: I0227 16:23:19.546126 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:19 crc kubenswrapper[4814]: I0227 16:23:19.664731 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:19 crc kubenswrapper[4814]: I0227 16:23:19.665541 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:19 crc kubenswrapper[4814]: I0227 16:23:19.666983 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:19 crc kubenswrapper[4814]: I0227 16:23:19.667065 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:19 crc kubenswrapper[4814]: I0227 16:23:19.667083 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:19 crc kubenswrapper[4814]: I0227 16:23:19.672591 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:20 crc kubenswrapper[4814]: I0227 16:23:20.414654 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 27 16:23:20 crc kubenswrapper[4814]: I0227 16:23:20.414856 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:20 crc kubenswrapper[4814]: I0227 16:23:20.416030 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:20 crc kubenswrapper[4814]: I0227 16:23:20.416067 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:20 crc kubenswrapper[4814]: I0227 16:23:20.416077 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:20 crc kubenswrapper[4814]: I0227 16:23:20.576200 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:20 crc kubenswrapper[4814]: I0227 16:23:20.577554 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:20 crc kubenswrapper[4814]: I0227 16:23:20.577667 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:20 crc kubenswrapper[4814]: I0227 16:23:20.577737 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:22 crc kubenswrapper[4814]: I0227 16:23:22.398193 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 27 16:23:22 crc kubenswrapper[4814]: W0227 16:23:22.557001 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 27 16:23:22 crc kubenswrapper[4814]: I0227 16:23:22.557147 4814 trace.go:236] Trace[2071457516]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Feb-2026 16:23:12.554) (total time: 10002ms): Feb 27 16:23:22 crc kubenswrapper[4814]: Trace[2071457516]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (16:23:22.556) Feb 27 16:23:22 crc kubenswrapper[4814]: Trace[2071457516]: [10.00220197s] [10.00220197s] END Feb 27 16:23:22 crc kubenswrapper[4814]: E0227 16:23:22.557176 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 27 16:23:22 crc kubenswrapper[4814]: I0227 16:23:22.665178 4814 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 16:23:22 crc kubenswrapper[4814]: I0227 16:23:22.665341 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 16:23:22 crc kubenswrapper[4814]: E0227 16:23:22.790533 4814 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:22Z is after 2026-02-23T05:33:13Z" node="crc" Feb 27 16:23:22 crc kubenswrapper[4814]: W0227 16:23:22.792003 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:22Z is after 2026-02-23T05:33:13Z Feb 27 16:23:22 crc kubenswrapper[4814]: E0227 16:23:22.792151 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:22Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 16:23:22 crc kubenswrapper[4814]: E0227 16:23:22.792325 4814 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:22Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.1898270c965ddbb2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.386114482 +0000 UTC m=+0.838739312,LastTimestamp:2026-02-27 16:23:08.386114482 +0000 UTC m=+0.838739312,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:23:22 crc kubenswrapper[4814]: E0227 16:23:22.794247 4814 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:22Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 16:23:22 crc kubenswrapper[4814]: E0227 16:23:22.796039 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:22Z is after 2026-02-23T05:33:13Z" interval="6.4s" Feb 27 16:23:22 crc kubenswrapper[4814]: W0227 16:23:22.797431 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:22Z is after 2026-02-23T05:33:13Z Feb 27 16:23:22 crc kubenswrapper[4814]: E0227 16:23:22.797528 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:22Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 16:23:22 crc kubenswrapper[4814]: W0227 16:23:22.799854 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:22Z is after 2026-02-23T05:33:13Z Feb 27 16:23:22 crc kubenswrapper[4814]: E0227 16:23:22.799937 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:22Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 16:23:22 crc kubenswrapper[4814]: I0227 16:23:22.806213 4814 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 27 16:23:22 crc kubenswrapper[4814]: I0227 16:23:22.806304 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 27 16:23:22 crc kubenswrapper[4814]: I0227 16:23:22.812509 4814 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 27 16:23:22 crc kubenswrapper[4814]: I0227 16:23:22.812587 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 27 16:23:23 crc kubenswrapper[4814]: I0227 16:23:23.401213 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:23Z is after 2026-02-23T05:33:13Z Feb 27 16:23:23 crc kubenswrapper[4814]: I0227 16:23:23.439975 4814 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 27 16:23:23 crc kubenswrapper[4814]: I0227 16:23:23.440090 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 27 16:23:23 crc kubenswrapper[4814]: I0227 16:23:23.475668 4814 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]log ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]etcd ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/openshift.io-api-request-count-filter ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/openshift.io-startkubeinformers ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/generic-apiserver-start-informers ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/priority-and-fairness-config-consumer ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/priority-and-fairness-filter ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/start-apiextensions-informers ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/start-apiextensions-controllers ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/crd-informer-synced ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/start-system-namespaces-controller ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/start-cluster-authentication-info-controller ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/start-legacy-token-tracking-controller ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/start-service-ip-repair-controllers ok Feb 27 16:23:23 crc kubenswrapper[4814]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Feb 27 16:23:23 crc kubenswrapper[4814]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/priority-and-fairness-config-producer ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/bootstrap-controller ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/start-kube-aggregator-informers ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/apiservice-status-local-available-controller ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/apiservice-status-remote-available-controller ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/apiservice-registration-controller ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/apiservice-wait-for-first-sync ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/apiservice-discovery-controller ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/kube-apiserver-autoregistration ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]autoregister-completion ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/apiservice-openapi-controller ok Feb 27 16:23:23 crc kubenswrapper[4814]: [+]poststarthook/apiservice-openapiv3-controller ok Feb 27 16:23:23 crc kubenswrapper[4814]: livez check failed Feb 27 16:23:23 crc kubenswrapper[4814]: I0227 16:23:23.475742 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:23:23 crc kubenswrapper[4814]: I0227 16:23:23.593103 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 27 16:23:23 crc kubenswrapper[4814]: I0227 16:23:23.596308 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"999cf79255061bae18e7033b5e5cbbe73b0f3b85421b580972e97ed99db6b363"} Feb 27 16:23:23 crc kubenswrapper[4814]: I0227 16:23:23.596497 4814 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="999cf79255061bae18e7033b5e5cbbe73b0f3b85421b580972e97ed99db6b363" exitCode=255 Feb 27 16:23:23 crc kubenswrapper[4814]: I0227 16:23:23.596732 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:23 crc kubenswrapper[4814]: I0227 16:23:23.598594 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:23 crc kubenswrapper[4814]: I0227 16:23:23.598625 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:23 crc kubenswrapper[4814]: I0227 16:23:23.598634 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:23 crc kubenswrapper[4814]: I0227 16:23:23.599101 4814 scope.go:117] "RemoveContainer" containerID="999cf79255061bae18e7033b5e5cbbe73b0f3b85421b580972e97ed99db6b363" Feb 27 16:23:24 crc kubenswrapper[4814]: I0227 16:23:24.405774 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:24Z is after 2026-02-23T05:33:13Z Feb 27 16:23:24 crc kubenswrapper[4814]: I0227 16:23:24.602452 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 27 16:23:24 crc kubenswrapper[4814]: I0227 16:23:24.604924 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6943c9e4955175193647881bcf453ef492352005e0132940c758394ba5f5d30e"} Feb 27 16:23:24 crc kubenswrapper[4814]: I0227 16:23:24.605089 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:24 crc kubenswrapper[4814]: I0227 16:23:24.606172 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:24 crc kubenswrapper[4814]: I0227 16:23:24.606202 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:24 crc kubenswrapper[4814]: I0227 16:23:24.606211 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:25 crc kubenswrapper[4814]: I0227 16:23:25.400833 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:25Z is after 2026-02-23T05:33:13Z Feb 27 16:23:25 crc kubenswrapper[4814]: I0227 16:23:25.610245 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 27 16:23:25 crc kubenswrapper[4814]: I0227 16:23:25.611006 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 27 16:23:25 crc kubenswrapper[4814]: I0227 16:23:25.613523 4814 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6943c9e4955175193647881bcf453ef492352005e0132940c758394ba5f5d30e" exitCode=255 Feb 27 16:23:25 crc kubenswrapper[4814]: I0227 16:23:25.613589 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6943c9e4955175193647881bcf453ef492352005e0132940c758394ba5f5d30e"} Feb 27 16:23:25 crc kubenswrapper[4814]: I0227 16:23:25.613669 4814 scope.go:117] "RemoveContainer" containerID="999cf79255061bae18e7033b5e5cbbe73b0f3b85421b580972e97ed99db6b363" Feb 27 16:23:25 crc kubenswrapper[4814]: I0227 16:23:25.613827 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:25 crc kubenswrapper[4814]: I0227 16:23:25.615005 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:25 crc kubenswrapper[4814]: I0227 16:23:25.615056 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:25 crc kubenswrapper[4814]: I0227 16:23:25.615107 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:25 crc kubenswrapper[4814]: I0227 16:23:25.616150 4814 scope.go:117] "RemoveContainer" containerID="6943c9e4955175193647881bcf453ef492352005e0132940c758394ba5f5d30e" Feb 27 16:23:25 crc kubenswrapper[4814]: E0227 16:23:25.616489 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 16:23:26 crc kubenswrapper[4814]: I0227 16:23:26.404618 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:26Z is after 2026-02-23T05:33:13Z Feb 27 16:23:26 crc kubenswrapper[4814]: I0227 16:23:26.620112 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 27 16:23:27 crc kubenswrapper[4814]: W0227 16:23:27.240231 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:27Z is after 2026-02-23T05:33:13Z Feb 27 16:23:27 crc kubenswrapper[4814]: E0227 16:23:27.240420 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:27Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 16:23:27 crc kubenswrapper[4814]: I0227 16:23:27.402077 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:27Z is after 2026-02-23T05:33:13Z Feb 27 16:23:28 crc kubenswrapper[4814]: I0227 16:23:28.402739 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:28Z is after 2026-02-23T05:33:13Z Feb 27 16:23:28 crc kubenswrapper[4814]: I0227 16:23:28.475127 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:28 crc kubenswrapper[4814]: I0227 16:23:28.475361 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:28 crc kubenswrapper[4814]: I0227 16:23:28.477393 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:28 crc kubenswrapper[4814]: I0227 16:23:28.477442 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:28 crc kubenswrapper[4814]: I0227 16:23:28.477453 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:28 crc kubenswrapper[4814]: I0227 16:23:28.478152 4814 scope.go:117] "RemoveContainer" containerID="6943c9e4955175193647881bcf453ef492352005e0132940c758394ba5f5d30e" Feb 27 16:23:28 crc kubenswrapper[4814]: E0227 16:23:28.478375 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 16:23:28 crc kubenswrapper[4814]: I0227 16:23:28.481598 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:28 crc kubenswrapper[4814]: E0227 16:23:28.575173 4814 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 16:23:28 crc kubenswrapper[4814]: I0227 16:23:28.628793 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:28 crc kubenswrapper[4814]: I0227 16:23:28.630548 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:28 crc kubenswrapper[4814]: I0227 16:23:28.630587 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:28 crc kubenswrapper[4814]: I0227 16:23:28.630602 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:28 crc kubenswrapper[4814]: I0227 16:23:28.631237 4814 scope.go:117] "RemoveContainer" containerID="6943c9e4955175193647881bcf453ef492352005e0132940c758394ba5f5d30e" Feb 27 16:23:28 crc kubenswrapper[4814]: E0227 16:23:28.631472 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 16:23:29 crc kubenswrapper[4814]: I0227 16:23:29.191062 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:29 crc kubenswrapper[4814]: I0227 16:23:29.192969 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:29 crc kubenswrapper[4814]: I0227 16:23:29.193051 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:29 crc kubenswrapper[4814]: I0227 16:23:29.193079 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:29 crc kubenswrapper[4814]: I0227 16:23:29.193119 4814 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 16:23:29 crc kubenswrapper[4814]: E0227 16:23:29.197684 4814 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:29Z is after 2026-02-23T05:33:13Z" node="crc" Feb 27 16:23:29 crc kubenswrapper[4814]: E0227 16:23:29.201045 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:29Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 27 16:23:29 crc kubenswrapper[4814]: I0227 16:23:29.400398 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:29Z is after 2026-02-23T05:33:13Z Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.279643 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.279930 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.281347 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.281385 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.281395 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.282078 4814 scope.go:117] "RemoveContainer" containerID="6943c9e4955175193647881bcf453ef492352005e0132940c758394ba5f5d30e" Feb 27 16:23:30 crc kubenswrapper[4814]: E0227 16:23:30.282280 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.402331 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:30Z is after 2026-02-23T05:33:13Z Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.444273 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.444922 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.446775 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.446830 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.446849 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.461785 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 27 16:23:30 crc kubenswrapper[4814]: W0227 16:23:30.463164 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:30Z is after 2026-02-23T05:33:13Z Feb 27 16:23:30 crc kubenswrapper[4814]: E0227 16:23:30.463291 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:30Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.634979 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.636029 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.636064 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:30 crc kubenswrapper[4814]: I0227 16:23:30.636076 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:31 crc kubenswrapper[4814]: I0227 16:23:31.152376 4814 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 16:23:31 crc kubenswrapper[4814]: E0227 16:23:31.157459 4814 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:31Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 16:23:31 crc kubenswrapper[4814]: W0227 16:23:31.195204 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:31Z is after 2026-02-23T05:33:13Z Feb 27 16:23:31 crc kubenswrapper[4814]: E0227 16:23:31.195330 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:31Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 16:23:31 crc kubenswrapper[4814]: I0227 16:23:31.400217 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:31Z is after 2026-02-23T05:33:13Z Feb 27 16:23:31 crc kubenswrapper[4814]: W0227 16:23:31.927070 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:31Z is after 2026-02-23T05:33:13Z Feb 27 16:23:31 crc kubenswrapper[4814]: E0227 16:23:31.927190 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:31Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 16:23:32 crc kubenswrapper[4814]: I0227 16:23:32.402531 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:32Z is after 2026-02-23T05:33:13Z Feb 27 16:23:32 crc kubenswrapper[4814]: I0227 16:23:32.665517 4814 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 16:23:32 crc kubenswrapper[4814]: I0227 16:23:32.665600 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 16:23:32 crc kubenswrapper[4814]: E0227 16:23:32.796107 4814 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:32Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.1898270c965ddbb2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.386114482 +0000 UTC m=+0.838739312,LastTimestamp:2026-02-27 16:23:08.386114482 +0000 UTC m=+0.838739312,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:23:33 crc kubenswrapper[4814]: I0227 16:23:33.402369 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:33Z is after 2026-02-23T05:33:13Z Feb 27 16:23:33 crc kubenswrapper[4814]: I0227 16:23:33.439800 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:33 crc kubenswrapper[4814]: I0227 16:23:33.440104 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:33 crc kubenswrapper[4814]: I0227 16:23:33.441942 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:33 crc kubenswrapper[4814]: I0227 16:23:33.442033 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:33 crc kubenswrapper[4814]: I0227 16:23:33.442048 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:33 crc kubenswrapper[4814]: I0227 16:23:33.442929 4814 scope.go:117] "RemoveContainer" containerID="6943c9e4955175193647881bcf453ef492352005e0132940c758394ba5f5d30e" Feb 27 16:23:33 crc kubenswrapper[4814]: E0227 16:23:33.443178 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 16:23:34 crc kubenswrapper[4814]: I0227 16:23:34.403562 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:34Z is after 2026-02-23T05:33:13Z Feb 27 16:23:35 crc kubenswrapper[4814]: I0227 16:23:35.402921 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:35Z is after 2026-02-23T05:33:13Z Feb 27 16:23:36 crc kubenswrapper[4814]: I0227 16:23:36.198735 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:36 crc kubenswrapper[4814]: I0227 16:23:36.200678 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:36 crc kubenswrapper[4814]: I0227 16:23:36.200734 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:36 crc kubenswrapper[4814]: I0227 16:23:36.200755 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:36 crc kubenswrapper[4814]: I0227 16:23:36.200788 4814 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 16:23:36 crc kubenswrapper[4814]: E0227 16:23:36.205941 4814 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:36Z is after 2026-02-23T05:33:13Z" node="crc" Feb 27 16:23:36 crc kubenswrapper[4814]: E0227 16:23:36.209393 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:36Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 27 16:23:36 crc kubenswrapper[4814]: I0227 16:23:36.402227 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:36Z is after 2026-02-23T05:33:13Z Feb 27 16:23:37 crc kubenswrapper[4814]: I0227 16:23:37.402432 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:37Z is after 2026-02-23T05:33:13Z Feb 27 16:23:38 crc kubenswrapper[4814]: I0227 16:23:38.400292 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:38Z is after 2026-02-23T05:33:13Z Feb 27 16:23:38 crc kubenswrapper[4814]: E0227 16:23:38.575343 4814 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 16:23:39 crc kubenswrapper[4814]: I0227 16:23:39.402053 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:39Z is after 2026-02-23T05:33:13Z Feb 27 16:23:39 crc kubenswrapper[4814]: W0227 16:23:39.429233 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:39Z is after 2026-02-23T05:33:13Z Feb 27 16:23:39 crc kubenswrapper[4814]: E0227 16:23:39.429387 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:39Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 16:23:40 crc kubenswrapper[4814]: I0227 16:23:40.401685 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:40Z is after 2026-02-23T05:33:13Z Feb 27 16:23:41 crc kubenswrapper[4814]: I0227 16:23:41.400131 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:41Z is after 2026-02-23T05:33:13Z Feb 27 16:23:41 crc kubenswrapper[4814]: I0227 16:23:41.476000 4814 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:36310->192.168.126.11:10357: read: connection reset by peer" start-of-body= Feb 27 16:23:41 crc kubenswrapper[4814]: I0227 16:23:41.476073 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:36310->192.168.126.11:10357: read: connection reset by peer" Feb 27 16:23:41 crc kubenswrapper[4814]: I0227 16:23:41.476138 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:41 crc kubenswrapper[4814]: I0227 16:23:41.476321 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:41 crc kubenswrapper[4814]: I0227 16:23:41.477621 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:41 crc kubenswrapper[4814]: I0227 16:23:41.477695 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:41 crc kubenswrapper[4814]: I0227 16:23:41.477715 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:41 crc kubenswrapper[4814]: I0227 16:23:41.478762 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"b5ba138e1d2cbeab8ba50d4d02b246c31cb1226615131fa4b074a08d615ca4f8"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Feb 27 16:23:41 crc kubenswrapper[4814]: I0227 16:23:41.479068 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://b5ba138e1d2cbeab8ba50d4d02b246c31cb1226615131fa4b074a08d615ca4f8" gracePeriod=30 Feb 27 16:23:41 crc kubenswrapper[4814]: I0227 16:23:41.668815 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 27 16:23:41 crc kubenswrapper[4814]: I0227 16:23:41.669597 4814 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="b5ba138e1d2cbeab8ba50d4d02b246c31cb1226615131fa4b074a08d615ca4f8" exitCode=255 Feb 27 16:23:41 crc kubenswrapper[4814]: I0227 16:23:41.669649 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"b5ba138e1d2cbeab8ba50d4d02b246c31cb1226615131fa4b074a08d615ca4f8"} Feb 27 16:23:42 crc kubenswrapper[4814]: I0227 16:23:42.400827 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:42Z is after 2026-02-23T05:33:13Z Feb 27 16:23:42 crc kubenswrapper[4814]: I0227 16:23:42.676070 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 27 16:23:42 crc kubenswrapper[4814]: I0227 16:23:42.676636 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a22d8635e425d164cd7c63685f0fb8b64a69919a2a687c9ca42321bfed7559ba"} Feb 27 16:23:42 crc kubenswrapper[4814]: I0227 16:23:42.676791 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:42 crc kubenswrapper[4814]: I0227 16:23:42.677844 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:42 crc kubenswrapper[4814]: I0227 16:23:42.677895 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:42 crc kubenswrapper[4814]: I0227 16:23:42.677912 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:42 crc kubenswrapper[4814]: E0227 16:23:42.801635 4814 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:42Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.1898270c965ddbb2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.386114482 +0000 UTC m=+0.838739312,LastTimestamp:2026-02-27 16:23:08.386114482 +0000 UTC m=+0.838739312,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:23:43 crc kubenswrapper[4814]: I0227 16:23:43.207099 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:43 crc kubenswrapper[4814]: I0227 16:23:43.208739 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:43 crc kubenswrapper[4814]: I0227 16:23:43.208815 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:43 crc kubenswrapper[4814]: I0227 16:23:43.208840 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:43 crc kubenswrapper[4814]: I0227 16:23:43.208890 4814 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 16:23:43 crc kubenswrapper[4814]: E0227 16:23:43.218420 4814 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:43Z is after 2026-02-23T05:33:13Z" node="crc" Feb 27 16:23:43 crc kubenswrapper[4814]: E0227 16:23:43.221751 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:43Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 27 16:23:43 crc kubenswrapper[4814]: I0227 16:23:43.402588 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:43Z is after 2026-02-23T05:33:13Z Feb 27 16:23:43 crc kubenswrapper[4814]: I0227 16:23:43.679782 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:43 crc kubenswrapper[4814]: I0227 16:23:43.681057 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:43 crc kubenswrapper[4814]: I0227 16:23:43.681112 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:43 crc kubenswrapper[4814]: I0227 16:23:43.681123 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:44 crc kubenswrapper[4814]: I0227 16:23:44.401309 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:44Z is after 2026-02-23T05:33:13Z Feb 27 16:23:44 crc kubenswrapper[4814]: W0227 16:23:44.764024 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:44Z is after 2026-02-23T05:33:13Z Feb 27 16:23:44 crc kubenswrapper[4814]: E0227 16:23:44.764124 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:44Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 16:23:45 crc kubenswrapper[4814]: I0227 16:23:45.404079 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:45Z is after 2026-02-23T05:33:13Z Feb 27 16:23:46 crc kubenswrapper[4814]: I0227 16:23:46.388545 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:46 crc kubenswrapper[4814]: I0227 16:23:46.388683 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:46 crc kubenswrapper[4814]: I0227 16:23:46.389731 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:46 crc kubenswrapper[4814]: I0227 16:23:46.389754 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:46 crc kubenswrapper[4814]: I0227 16:23:46.389763 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:46 crc kubenswrapper[4814]: I0227 16:23:46.402444 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:46Z is after 2026-02-23T05:33:13Z Feb 27 16:23:47 crc kubenswrapper[4814]: I0227 16:23:47.403536 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:47Z is after 2026-02-23T05:33:13Z Feb 27 16:23:47 crc kubenswrapper[4814]: I0227 16:23:47.487587 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:47 crc kubenswrapper[4814]: I0227 16:23:47.489058 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:47 crc kubenswrapper[4814]: I0227 16:23:47.489144 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:47 crc kubenswrapper[4814]: I0227 16:23:47.489205 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:47 crc kubenswrapper[4814]: I0227 16:23:47.490092 4814 scope.go:117] "RemoveContainer" containerID="6943c9e4955175193647881bcf453ef492352005e0132940c758394ba5f5d30e" Feb 27 16:23:47 crc kubenswrapper[4814]: I0227 16:23:47.786856 4814 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 16:23:47 crc kubenswrapper[4814]: E0227 16:23:47.792866 4814 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:47Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 16:23:47 crc kubenswrapper[4814]: E0227 16:23:47.794066 4814 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Feb 27 16:23:48 crc kubenswrapper[4814]: I0227 16:23:48.402346 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:48Z is after 2026-02-23T05:33:13Z Feb 27 16:23:48 crc kubenswrapper[4814]: E0227 16:23:48.575498 4814 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 16:23:48 crc kubenswrapper[4814]: I0227 16:23:48.698308 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 27 16:23:48 crc kubenswrapper[4814]: I0227 16:23:48.701433 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fbc8df752c45f857b0470acc6fa6ebd84ca0eeb96a33a6fdb790bfd28df3f07c"} Feb 27 16:23:48 crc kubenswrapper[4814]: I0227 16:23:48.701626 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:48 crc kubenswrapper[4814]: I0227 16:23:48.702881 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:48 crc kubenswrapper[4814]: I0227 16:23:48.702939 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:48 crc kubenswrapper[4814]: I0227 16:23:48.702958 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:48 crc kubenswrapper[4814]: W0227 16:23:48.796047 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:48Z is after 2026-02-23T05:33:13Z Feb 27 16:23:48 crc kubenswrapper[4814]: E0227 16:23:48.796158 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:48Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 16:23:49 crc kubenswrapper[4814]: I0227 16:23:49.405612 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:49Z is after 2026-02-23T05:33:13Z Feb 27 16:23:49 crc kubenswrapper[4814]: I0227 16:23:49.664843 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:23:49 crc kubenswrapper[4814]: I0227 16:23:49.665672 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:49 crc kubenswrapper[4814]: I0227 16:23:49.667479 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:49 crc kubenswrapper[4814]: I0227 16:23:49.667539 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:49 crc kubenswrapper[4814]: I0227 16:23:49.667560 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:49 crc kubenswrapper[4814]: I0227 16:23:49.706968 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 27 16:23:49 crc kubenswrapper[4814]: I0227 16:23:49.707566 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 27 16:23:49 crc kubenswrapper[4814]: I0227 16:23:49.710447 4814 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fbc8df752c45f857b0470acc6fa6ebd84ca0eeb96a33a6fdb790bfd28df3f07c" exitCode=255 Feb 27 16:23:49 crc kubenswrapper[4814]: I0227 16:23:49.710494 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"fbc8df752c45f857b0470acc6fa6ebd84ca0eeb96a33a6fdb790bfd28df3f07c"} Feb 27 16:23:49 crc kubenswrapper[4814]: I0227 16:23:49.710572 4814 scope.go:117] "RemoveContainer" containerID="6943c9e4955175193647881bcf453ef492352005e0132940c758394ba5f5d30e" Feb 27 16:23:49 crc kubenswrapper[4814]: I0227 16:23:49.710709 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:49 crc kubenswrapper[4814]: I0227 16:23:49.711951 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:49 crc kubenswrapper[4814]: I0227 16:23:49.711987 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:49 crc kubenswrapper[4814]: I0227 16:23:49.711996 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:49 crc kubenswrapper[4814]: I0227 16:23:49.712563 4814 scope.go:117] "RemoveContainer" containerID="fbc8df752c45f857b0470acc6fa6ebd84ca0eeb96a33a6fdb790bfd28df3f07c" Feb 27 16:23:49 crc kubenswrapper[4814]: E0227 16:23:49.712795 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 16:23:50 crc kubenswrapper[4814]: I0227 16:23:50.218897 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:50 crc kubenswrapper[4814]: I0227 16:23:50.220731 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:50 crc kubenswrapper[4814]: I0227 16:23:50.220779 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:50 crc kubenswrapper[4814]: I0227 16:23:50.220797 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:50 crc kubenswrapper[4814]: I0227 16:23:50.220828 4814 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 16:23:50 crc kubenswrapper[4814]: E0227 16:23:50.226104 4814 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:50Z is after 2026-02-23T05:33:13Z" node="crc" Feb 27 16:23:50 crc kubenswrapper[4814]: E0227 16:23:50.229786 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:50Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 27 16:23:50 crc kubenswrapper[4814]: I0227 16:23:50.279212 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:50 crc kubenswrapper[4814]: I0227 16:23:50.401363 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:50Z is after 2026-02-23T05:33:13Z Feb 27 16:23:50 crc kubenswrapper[4814]: I0227 16:23:50.715804 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 27 16:23:50 crc kubenswrapper[4814]: I0227 16:23:50.719345 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:50 crc kubenswrapper[4814]: I0227 16:23:50.720729 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:50 crc kubenswrapper[4814]: I0227 16:23:50.720791 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:50 crc kubenswrapper[4814]: I0227 16:23:50.720803 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:50 crc kubenswrapper[4814]: I0227 16:23:50.721390 4814 scope.go:117] "RemoveContainer" containerID="fbc8df752c45f857b0470acc6fa6ebd84ca0eeb96a33a6fdb790bfd28df3f07c" Feb 27 16:23:50 crc kubenswrapper[4814]: E0227 16:23:50.721589 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 16:23:51 crc kubenswrapper[4814]: I0227 16:23:51.403211 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:51Z is after 2026-02-23T05:33:13Z Feb 27 16:23:52 crc kubenswrapper[4814]: I0227 16:23:52.401160 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:52Z is after 2026-02-23T05:33:13Z Feb 27 16:23:52 crc kubenswrapper[4814]: I0227 16:23:52.665683 4814 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 16:23:52 crc kubenswrapper[4814]: I0227 16:23:52.665784 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 16:23:52 crc kubenswrapper[4814]: E0227 16:23:52.807445 4814 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:52Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.1898270c965ddbb2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.386114482 +0000 UTC m=+0.838739312,LastTimestamp:2026-02-27 16:23:08.386114482 +0000 UTC m=+0.838739312,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:23:53 crc kubenswrapper[4814]: I0227 16:23:53.403085 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:53Z is after 2026-02-23T05:33:13Z Feb 27 16:23:53 crc kubenswrapper[4814]: I0227 16:23:53.439306 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:23:53 crc kubenswrapper[4814]: I0227 16:23:53.439551 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:53 crc kubenswrapper[4814]: I0227 16:23:53.441433 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:53 crc kubenswrapper[4814]: I0227 16:23:53.441485 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:53 crc kubenswrapper[4814]: I0227 16:23:53.441501 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:53 crc kubenswrapper[4814]: I0227 16:23:53.442453 4814 scope.go:117] "RemoveContainer" containerID="fbc8df752c45f857b0470acc6fa6ebd84ca0eeb96a33a6fdb790bfd28df3f07c" Feb 27 16:23:53 crc kubenswrapper[4814]: E0227 16:23:53.442729 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 16:23:54 crc kubenswrapper[4814]: I0227 16:23:54.402505 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:54Z is after 2026-02-23T05:33:13Z Feb 27 16:23:55 crc kubenswrapper[4814]: I0227 16:23:55.402175 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:55Z is after 2026-02-23T05:33:13Z Feb 27 16:23:56 crc kubenswrapper[4814]: W0227 16:23:56.317497 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:56Z is after 2026-02-23T05:33:13Z Feb 27 16:23:56 crc kubenswrapper[4814]: E0227 16:23:56.317606 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:56Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 16:23:56 crc kubenswrapper[4814]: I0227 16:23:56.402568 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:56Z is after 2026-02-23T05:33:13Z Feb 27 16:23:57 crc kubenswrapper[4814]: I0227 16:23:57.226415 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:23:57 crc kubenswrapper[4814]: I0227 16:23:57.227983 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:23:57 crc kubenswrapper[4814]: I0227 16:23:57.228029 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:23:57 crc kubenswrapper[4814]: I0227 16:23:57.228042 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:23:57 crc kubenswrapper[4814]: I0227 16:23:57.228068 4814 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 16:23:57 crc kubenswrapper[4814]: E0227 16:23:57.232580 4814 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:57Z is after 2026-02-23T05:33:13Z" node="crc" Feb 27 16:23:57 crc kubenswrapper[4814]: E0227 16:23:57.236380 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:57Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 27 16:23:57 crc kubenswrapper[4814]: I0227 16:23:57.402350 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:57Z is after 2026-02-23T05:33:13Z Feb 27 16:23:58 crc kubenswrapper[4814]: I0227 16:23:58.402804 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:58Z is after 2026-02-23T05:33:13Z Feb 27 16:23:58 crc kubenswrapper[4814]: E0227 16:23:58.575954 4814 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 16:23:59 crc kubenswrapper[4814]: I0227 16:23:59.402435 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:59Z is after 2026-02-23T05:33:13Z Feb 27 16:24:00 crc kubenswrapper[4814]: I0227 16:24:00.403056 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:00Z is after 2026-02-23T05:33:13Z Feb 27 16:24:01 crc kubenswrapper[4814]: I0227 16:24:01.400053 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:01Z is after 2026-02-23T05:33:13Z Feb 27 16:24:02 crc kubenswrapper[4814]: I0227 16:24:02.400187 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:02Z is after 2026-02-23T05:33:13Z Feb 27 16:24:02 crc kubenswrapper[4814]: I0227 16:24:02.664976 4814 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 16:24:02 crc kubenswrapper[4814]: I0227 16:24:02.665058 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.815726 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c965ddbb2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.386114482 +0000 UTC m=+0.838739312,LastTimestamp:2026-02-27 16:23:08.386114482 +0000 UTC m=+0.838739312,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.822564 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3b422e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467733038 +0000 UTC m=+0.920357888,LastTimestamp:2026-02-27 16:23:08.467733038 +0000 UTC m=+0.920357888,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.832285 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3ba1bb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467757499 +0000 UTC m=+0.920382329,LastTimestamp:2026-02-27 16:23:08.467757499 +0000 UTC m=+0.920382329,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.839426 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3bc9f7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467767799 +0000 UTC m=+0.920392639,LastTimestamp:2026-02-27 16:23:08.467767799 +0000 UTC m=+0.920392639,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.847318 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270ca1625c36 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.570958902 +0000 UTC m=+1.023583742,LastTimestamp:2026-02-27 16:23:08.570958902 +0000 UTC m=+1.023583742,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.855618 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3b422e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3b422e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467733038 +0000 UTC m=+0.920357888,LastTimestamp:2026-02-27 16:23:08.589030552 +0000 UTC m=+1.041655382,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.867688 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3ba1bb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3ba1bb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467757499 +0000 UTC m=+0.920382329,LastTimestamp:2026-02-27 16:23:08.589044462 +0000 UTC m=+1.041669282,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.874663 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3bc9f7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3bc9f7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467767799 +0000 UTC m=+0.920392639,LastTimestamp:2026-02-27 16:23:08.589054272 +0000 UTC m=+1.041679102,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.879592 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3b422e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3b422e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467733038 +0000 UTC m=+0.920357888,LastTimestamp:2026-02-27 16:23:08.590154249 +0000 UTC m=+1.042779119,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.884853 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3ba1bb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3ba1bb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467757499 +0000 UTC m=+0.920382329,LastTimestamp:2026-02-27 16:23:08.5901861 +0000 UTC m=+1.042810970,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.892448 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3bc9f7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3bc9f7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467767799 +0000 UTC m=+0.920392639,LastTimestamp:2026-02-27 16:23:08.59020185 +0000 UTC m=+1.042826710,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.898162 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3b422e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3b422e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467733038 +0000 UTC m=+0.920357888,LastTimestamp:2026-02-27 16:23:08.595305027 +0000 UTC m=+1.047929867,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.907044 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3ba1bb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3ba1bb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467757499 +0000 UTC m=+0.920382329,LastTimestamp:2026-02-27 16:23:08.595320827 +0000 UTC m=+1.047945667,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.909456 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3bc9f7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3bc9f7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467767799 +0000 UTC m=+0.920392639,LastTimestamp:2026-02-27 16:23:08.595331848 +0000 UTC m=+1.047956688,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.911812 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3b422e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3b422e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467733038 +0000 UTC m=+0.920357888,LastTimestamp:2026-02-27 16:23:08.59540805 +0000 UTC m=+1.048032920,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.915838 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3ba1bb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3ba1bb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467757499 +0000 UTC m=+0.920382329,LastTimestamp:2026-02-27 16:23:08.595427811 +0000 UTC m=+1.048052681,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.920798 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3bc9f7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3bc9f7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467767799 +0000 UTC m=+0.920392639,LastTimestamp:2026-02-27 16:23:08.595442051 +0000 UTC m=+1.048066921,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.927183 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3b422e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3b422e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467733038 +0000 UTC m=+0.920357888,LastTimestamp:2026-02-27 16:23:08.59631405 +0000 UTC m=+1.048938890,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.931866 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3ba1bb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3ba1bb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467757499 +0000 UTC m=+0.920382329,LastTimestamp:2026-02-27 16:23:08.596327641 +0000 UTC m=+1.048952481,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.936582 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3bc9f7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3bc9f7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467767799 +0000 UTC m=+0.920392639,LastTimestamp:2026-02-27 16:23:08.596365631 +0000 UTC m=+1.048990471,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.942586 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3b422e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3b422e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467733038 +0000 UTC m=+0.920357888,LastTimestamp:2026-02-27 16:23:08.597730456 +0000 UTC m=+1.050355286,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.946808 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3ba1bb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3ba1bb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467757499 +0000 UTC m=+0.920382329,LastTimestamp:2026-02-27 16:23:08.5978603 +0000 UTC m=+1.050485120,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.951683 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3bc9f7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3bc9f7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467767799 +0000 UTC m=+0.920392639,LastTimestamp:2026-02-27 16:23:08.597972794 +0000 UTC m=+1.050597624,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.957120 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3b422e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3b422e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467733038 +0000 UTC m=+0.920357888,LastTimestamp:2026-02-27 16:23:08.599813284 +0000 UTC m=+1.052438124,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.961900 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.1898270c9b3ba1bb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898270c9b3ba1bb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:08.467757499 +0000 UTC m=+0.920382329,LastTimestamp:2026-02-27 16:23:08.599882286 +0000 UTC m=+1.052507116,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.968322 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270cbd414619 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.038552601 +0000 UTC m=+1.491177431,LastTimestamp:2026-02-27 16:23:09.038552601 +0000 UTC m=+1.491177431,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.972873 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.1898270cbd4163f1 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.038560241 +0000 UTC m=+1.491185111,LastTimestamp:2026-02-27 16:23:09.038560241 +0000 UTC m=+1.491185111,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.979961 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270cbd4a8dae openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.03916075 +0000 UTC m=+1.491785580,LastTimestamp:2026-02-27 16:23:09.03916075 +0000 UTC m=+1.491785580,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.984877 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898270cbed28176 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.064847734 +0000 UTC m=+1.517472564,LastTimestamp:2026-02-27 16:23:09.064847734 +0000 UTC m=+1.517472564,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.989759 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270cbed6ce39 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.065129529 +0000 UTC m=+1.517754359,LastTimestamp:2026-02-27 16:23:09.065129529 +0000 UTC m=+1.517754359,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.994912 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.1898270ce4b48193 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.700415891 +0000 UTC m=+2.153040721,LastTimestamp:2026-02-27 16:23:09.700415891 +0000 UTC m=+2.153040721,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:02 crc kubenswrapper[4814]: E0227 16:24:02.999123 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270ce4b80aab openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.700647595 +0000 UTC m=+2.153272465,LastTimestamp:2026-02-27 16:23:09.700647595 +0000 UTC m=+2.153272465,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.003447 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898270ce5540635 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.710870069 +0000 UTC m=+2.163494899,LastTimestamp:2026-02-27 16:23:09.710870069 +0000 UTC m=+2.163494899,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.012570 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270ce5635560 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.711873376 +0000 UTC m=+2.164498206,LastTimestamp:2026-02-27 16:23:09.711873376 +0000 UTC m=+2.164498206,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.015320 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.1898270ce5780cb4 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.713231028 +0000 UTC m=+2.165855858,LastTimestamp:2026-02-27 16:23:09.713231028 +0000 UTC m=+2.165855858,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.022456 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270ce582e63f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.713942079 +0000 UTC m=+2.166566909,LastTimestamp:2026-02-27 16:23:09.713942079 +0000 UTC m=+2.166566909,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.029648 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270ce5991459 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.715395673 +0000 UTC m=+2.168020523,LastTimestamp:2026-02-27 16:23:09.715395673 +0000 UTC m=+2.168020523,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.036557 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270ce5c9f55d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.718599005 +0000 UTC m=+2.171223855,LastTimestamp:2026-02-27 16:23:09.718599005 +0000 UTC m=+2.171223855,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.042987 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898270ce615ad58 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.723561304 +0000 UTC m=+2.176186144,LastTimestamp:2026-02-27 16:23:09.723561304 +0000 UTC m=+2.176186144,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.047210 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270ce637a800 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.72578816 +0000 UTC m=+2.178412990,LastTimestamp:2026-02-27 16:23:09.72578816 +0000 UTC m=+2.178412990,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.051682 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270ce6f537db openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.738211291 +0000 UTC m=+2.190836121,LastTimestamp:2026-02-27 16:23:09.738211291 +0000 UTC m=+2.190836121,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.061668 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270cf719a684 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.009034372 +0000 UTC m=+2.461659242,LastTimestamp:2026-02-27 16:23:10.009034372 +0000 UTC m=+2.461659242,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.065828 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270cf7dc3d93 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.021787027 +0000 UTC m=+2.474411857,LastTimestamp:2026-02-27 16:23:10.021787027 +0000 UTC m=+2.474411857,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.071854 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270cf7f17116 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.02317647 +0000 UTC m=+2.475801300,LastTimestamp:2026-02-27 16:23:10.02317647 +0000 UTC m=+2.475801300,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.075993 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270d04d12d8f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.239165839 +0000 UTC m=+2.691790669,LastTimestamp:2026-02-27 16:23:10.239165839 +0000 UTC m=+2.691790669,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.079418 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270d059e45d8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.252606936 +0000 UTC m=+2.705231806,LastTimestamp:2026-02-27 16:23:10.252606936 +0000 UTC m=+2.705231806,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.083128 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270d05bc0d67 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.254558567 +0000 UTC m=+2.707183397,LastTimestamp:2026-02-27 16:23:10.254558567 +0000 UTC m=+2.707183397,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.088952 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270d12d93e0b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.474575371 +0000 UTC m=+2.927200201,LastTimestamp:2026-02-27 16:23:10.474575371 +0000 UTC m=+2.927200201,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.094864 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270d13d5ea04 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.491134468 +0000 UTC m=+2.943759328,LastTimestamp:2026-02-27 16:23:10.491134468 +0000 UTC m=+2.943759328,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.099001 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898270d14b70dfb openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.505889275 +0000 UTC m=+2.958514145,LastTimestamp:2026-02-27 16:23:10.505889275 +0000 UTC m=+2.958514145,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.104743 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270d15332ae0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.514023136 +0000 UTC m=+2.966647986,LastTimestamp:2026-02-27 16:23:10.514023136 +0000 UTC m=+2.966647986,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.109681 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270d1585aaa2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.519429794 +0000 UTC m=+2.972054664,LastTimestamp:2026-02-27 16:23:10.519429794 +0000 UTC m=+2.972054664,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.113595 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.1898270d158681d4 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.519484884 +0000 UTC m=+2.972109734,LastTimestamp:2026-02-27 16:23:10.519484884 +0000 UTC m=+2.972109734,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.117296 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898270d230234de openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.745695454 +0000 UTC m=+3.198320284,LastTimestamp:2026-02-27 16:23:10.745695454 +0000 UTC m=+3.198320284,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.122088 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898270d23ea93f4 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.760924148 +0000 UTC m=+3.213548978,LastTimestamp:2026-02-27 16:23:10.760924148 +0000 UTC m=+3.213548978,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.126579 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898270d23ff4072 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.762279026 +0000 UTC m=+3.214903856,LastTimestamp:2026-02-27 16:23:10.762279026 +0000 UTC m=+3.214903856,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.130419 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.1898270d245bf515 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.768354581 +0000 UTC m=+3.220979411,LastTimestamp:2026-02-27 16:23:10.768354581 +0000 UTC m=+3.220979411,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.136636 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270d246da36e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.769513326 +0000 UTC m=+3.222138156,LastTimestamp:2026-02-27 16:23:10.769513326 +0000 UTC m=+3.222138156,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.140708 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270d24f0816a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.778089834 +0000 UTC m=+3.230714664,LastTimestamp:2026-02-27 16:23:10.778089834 +0000 UTC m=+3.230714664,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.145414 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.1898270d25c15c29 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.791777321 +0000 UTC m=+3.244402151,LastTimestamp:2026-02-27 16:23:10.791777321 +0000 UTC m=+3.244402151,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.149107 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270d26308712 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.799062802 +0000 UTC m=+3.251687632,LastTimestamp:2026-02-27 16:23:10.799062802 +0000 UTC m=+3.251687632,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.152891 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270d2637b3af openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.799532975 +0000 UTC m=+3.252157805,LastTimestamp:2026-02-27 16:23:10.799532975 +0000 UTC m=+3.252157805,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.157019 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270d2648c547 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.800651591 +0000 UTC m=+3.253276421,LastTimestamp:2026-02-27 16:23:10.800651591 +0000 UTC m=+3.253276421,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.163028 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898270d309d6376 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.97396927 +0000 UTC m=+3.426594090,LastTimestamp:2026-02-27 16:23:10.97396927 +0000 UTC m=+3.426594090,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.167288 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898270d32513219 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.002530329 +0000 UTC m=+3.455155159,LastTimestamp:2026-02-27 16:23:11.002530329 +0000 UTC m=+3.455155159,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.173351 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898270d3266f737 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.003957047 +0000 UTC m=+3.456581917,LastTimestamp:2026-02-27 16:23:11.003957047 +0000 UTC m=+3.456581917,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.178747 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270d32c3a67f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.010031231 +0000 UTC m=+3.462656101,LastTimestamp:2026-02-27 16:23:11.010031231 +0000 UTC m=+3.462656101,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.184578 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270d348458a6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.039436966 +0000 UTC m=+3.492061826,LastTimestamp:2026-02-27 16:23:11.039436966 +0000 UTC m=+3.492061826,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.190914 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270d349f28e7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.041194215 +0000 UTC m=+3.493819075,LastTimestamp:2026-02-27 16:23:11.041194215 +0000 UTC m=+3.493819075,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.197075 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898270d3f96d83e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.225198654 +0000 UTC m=+3.677823494,LastTimestamp:2026-02-27 16:23:11.225198654 +0000 UTC m=+3.677823494,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.205843 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898270d40de7abd openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.246670525 +0000 UTC m=+3.699295365,LastTimestamp:2026-02-27 16:23:11.246670525 +0000 UTC m=+3.699295365,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.210439 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270d4269e82c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.27258526 +0000 UTC m=+3.725210130,LastTimestamp:2026-02-27 16:23:11.27258526 +0000 UTC m=+3.725210130,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.216744 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270d43564cd6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.288077526 +0000 UTC m=+3.740702396,LastTimestamp:2026-02-27 16:23:11.288077526 +0000 UTC m=+3.740702396,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.222116 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270d436c5efc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.289523964 +0000 UTC m=+3.742148794,LastTimestamp:2026-02-27 16:23:11.289523964 +0000 UTC m=+3.742148794,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.223977 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270d51bb7a31 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.529589297 +0000 UTC m=+3.982214127,LastTimestamp:2026-02-27 16:23:11.529589297 +0000 UTC m=+3.982214127,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.228112 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270d51df5ee0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.5319416 +0000 UTC m=+3.984566430,LastTimestamp:2026-02-27 16:23:11.5319416 +0000 UTC m=+3.984566430,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.235028 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270d52f10d55 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.549877589 +0000 UTC m=+4.002502419,LastTimestamp:2026-02-27 16:23:11.549877589 +0000 UTC m=+4.002502419,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.240650 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270d53027f9c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.551020956 +0000 UTC m=+4.003645806,LastTimestamp:2026-02-27 16:23:11.551020956 +0000 UTC m=+4.003645806,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.246716 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270d5de3e8f7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.733565687 +0000 UTC m=+4.186190517,LastTimestamp:2026-02-27 16:23:11.733565687 +0000 UTC m=+4.186190517,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.253662 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270d5df21cb0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.734496432 +0000 UTC m=+4.187121262,LastTimestamp:2026-02-27 16:23:11.734496432 +0000 UTC m=+4.187121262,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.260012 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270d5e8e41e1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.744729569 +0000 UTC m=+4.197354439,LastTimestamp:2026-02-27 16:23:11.744729569 +0000 UTC m=+4.197354439,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.266107 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270d5ed48661 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:11.749334625 +0000 UTC m=+4.201959455,LastTimestamp:2026-02-27 16:23:11.749334625 +0000 UTC m=+4.201959455,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.271470 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270d8e273700 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:12.543282944 +0000 UTC m=+4.995907774,LastTimestamp:2026-02-27 16:23:12.543282944 +0000 UTC m=+4.995907774,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.279125 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270d9b16ac37 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:12.760302647 +0000 UTC m=+5.212927507,LastTimestamp:2026-02-27 16:23:12.760302647 +0000 UTC m=+5.212927507,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.285137 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270d9bc46015 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:12.771686421 +0000 UTC m=+5.224311251,LastTimestamp:2026-02-27 16:23:12.771686421 +0000 UTC m=+5.224311251,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.291341 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270d9bd5efda openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:12.772837338 +0000 UTC m=+5.225462198,LastTimestamp:2026-02-27 16:23:12.772837338 +0000 UTC m=+5.225462198,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.297625 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270daa259a24 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:13.0129393 +0000 UTC m=+5.465564160,LastTimestamp:2026-02-27 16:23:13.0129393 +0000 UTC m=+5.465564160,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.304498 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270dab0cba20 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:13.028086304 +0000 UTC m=+5.480711134,LastTimestamp:2026-02-27 16:23:13.028086304 +0000 UTC m=+5.480711134,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.311302 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270dab201c76 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:13.029356662 +0000 UTC m=+5.481981492,LastTimestamp:2026-02-27 16:23:13.029356662 +0000 UTC m=+5.481981492,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.318206 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270db84785db openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:13.250043355 +0000 UTC m=+5.702668235,LastTimestamp:2026-02-27 16:23:13.250043355 +0000 UTC m=+5.702668235,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.324541 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270db93afddc openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:13.265999324 +0000 UTC m=+5.718624154,LastTimestamp:2026-02-27 16:23:13.265999324 +0000 UTC m=+5.718624154,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.330582 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270db94c6c9f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:13.267141791 +0000 UTC m=+5.719766661,LastTimestamp:2026-02-27 16:23:13.267141791 +0000 UTC m=+5.719766661,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.336569 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270dc6f3775e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:13.496192862 +0000 UTC m=+5.948817732,LastTimestamp:2026-02-27 16:23:13.496192862 +0000 UTC m=+5.948817732,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.341376 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270dc802bef1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:13.513971441 +0000 UTC m=+5.966596311,LastTimestamp:2026-02-27 16:23:13.513971441 +0000 UTC m=+5.966596311,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.347827 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270dc81894a9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:13.515402409 +0000 UTC m=+5.968027279,LastTimestamp:2026-02-27 16:23:13.515402409 +0000 UTC m=+5.968027279,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.353990 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270dd4fc3134 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:13.731645748 +0000 UTC m=+6.184270578,LastTimestamp:2026-02-27 16:23:13.731645748 +0000 UTC m=+6.184270578,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.358807 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898270dd60d91c0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:13.749561792 +0000 UTC m=+6.202186662,LastTimestamp:2026-02-27 16:23:13.749561792 +0000 UTC m=+6.202186662,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.379123 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 16:24:03 crc kubenswrapper[4814]: &Event{ObjectMeta:{kube-controller-manager-crc.1898270fe9792d96 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 27 16:24:03 crc kubenswrapper[4814]: body: Feb 27 16:24:03 crc kubenswrapper[4814]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:22.665315734 +0000 UTC m=+15.117940624,LastTimestamp:2026-02-27 16:23:22.665315734 +0000 UTC m=+15.117940624,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 16:24:03 crc kubenswrapper[4814]: > Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.385358 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270fe97a4228 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:22.665386536 +0000 UTC m=+15.118011406,LastTimestamp:2026-02-27 16:23:22.665386536 +0000 UTC m=+15.118011406,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.392179 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 27 16:24:03 crc kubenswrapper[4814]: &Event{ObjectMeta:{kube-apiserver-crc.1898270ff1e030fa openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Feb 27 16:24:03 crc kubenswrapper[4814]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 27 16:24:03 crc kubenswrapper[4814]: Feb 27 16:24:03 crc kubenswrapper[4814]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:22.806284538 +0000 UTC m=+15.258909408,LastTimestamp:2026-02-27 16:23:22.806284538 +0000 UTC m=+15.258909408,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 16:24:03 crc kubenswrapper[4814]: > Feb 27 16:24:03 crc kubenswrapper[4814]: I0227 16:24:03.399759 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.399839 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270ff1e0e91f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:22.806331679 +0000 UTC m=+15.258956539,LastTimestamp:2026-02-27 16:23:22.806331679 +0000 UTC m=+15.258956539,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.410142 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1898270ff1e030fa\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 27 16:24:03 crc kubenswrapper[4814]: &Event{ObjectMeta:{kube-apiserver-crc.1898270ff1e030fa openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Feb 27 16:24:03 crc kubenswrapper[4814]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 27 16:24:03 crc kubenswrapper[4814]: Feb 27 16:24:03 crc kubenswrapper[4814]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:22.806284538 +0000 UTC m=+15.258909408,LastTimestamp:2026-02-27 16:23:22.812564092 +0000 UTC m=+15.265188932,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 16:24:03 crc kubenswrapper[4814]: > Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.418998 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1898270ff1e0e91f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898270ff1e0e91f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:22.806331679 +0000 UTC m=+15.258956539,LastTimestamp:2026-02-27 16:23:22.812643504 +0000 UTC m=+15.265268344,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.423919 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 27 16:24:03 crc kubenswrapper[4814]: &Event{ObjectMeta:{kube-apiserver-crc.1898271017a6ba3f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Liveness probe error: Get "https://192.168.126.11:17697/healthz": dial tcp 192.168.126.11:17697: connect: connection refused Feb 27 16:24:03 crc kubenswrapper[4814]: body: Feb 27 16:24:03 crc kubenswrapper[4814]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:23.440052799 +0000 UTC m=+15.892677669,LastTimestamp:2026-02-27 16:23:23.440052799 +0000 UTC m=+15.892677669,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 16:24:03 crc kubenswrapper[4814]: > Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.429359 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898271017a7eb4c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Liveness probe failed: Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:23.440130892 +0000 UTC m=+15.892755772,LastTimestamp:2026-02-27 16:23:23.440130892 +0000 UTC m=+15.892755772,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.434688 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 27 16:24:03 crc kubenswrapper[4814]: &Event{ObjectMeta:{kube-apiserver-crc.1898271019c7104d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 500 Feb 27 16:24:03 crc kubenswrapper[4814]: body: [+]ping ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]log ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]etcd ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/openshift.io-api-request-count-filter ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/openshift.io-startkubeinformers ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/generic-apiserver-start-informers ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/priority-and-fairness-config-consumer ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/priority-and-fairness-filter ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/start-apiextensions-informers ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/start-apiextensions-controllers ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/crd-informer-synced ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/start-system-namespaces-controller ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/start-cluster-authentication-info-controller ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/start-legacy-token-tracking-controller ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/start-service-ip-repair-controllers ok Feb 27 16:24:03 crc kubenswrapper[4814]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Feb 27 16:24:03 crc kubenswrapper[4814]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/priority-and-fairness-config-producer ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/bootstrap-controller ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/start-kube-aggregator-informers ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/apiservice-status-local-available-controller ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/apiservice-status-remote-available-controller ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/apiservice-registration-controller ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/apiservice-wait-for-first-sync ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/apiservice-discovery-controller ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/kube-apiserver-autoregistration ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]autoregister-completion ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/apiservice-openapi-controller ok Feb 27 16:24:03 crc kubenswrapper[4814]: [+]poststarthook/apiservice-openapiv3-controller ok Feb 27 16:24:03 crc kubenswrapper[4814]: livez check failed Feb 27 16:24:03 crc kubenswrapper[4814]: Feb 27 16:24:03 crc kubenswrapper[4814]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:23.475726413 +0000 UTC m=+15.928351253,LastTimestamp:2026-02-27 16:23:23.475726413 +0000 UTC m=+15.928351253,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 16:24:03 crc kubenswrapper[4814]: > Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.443715 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.1898270fe9792d96\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 16:24:03 crc kubenswrapper[4814]: &Event{ObjectMeta:{kube-controller-manager-crc.1898270fe9792d96 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 27 16:24:03 crc kubenswrapper[4814]: body: Feb 27 16:24:03 crc kubenswrapper[4814]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:22.665315734 +0000 UTC m=+15.117940624,LastTimestamp:2026-02-27 16:23:32.66557698 +0000 UTC m=+25.118201840,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 16:24:03 crc kubenswrapper[4814]: > Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.450084 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.1898270fe97a4228\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270fe97a4228 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:22.665386536 +0000 UTC m=+15.118011406,LastTimestamp:2026-02-27 16:23:32.665634122 +0000 UTC m=+25.118258992,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.457339 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 16:24:03 crc kubenswrapper[4814]: &Event{ObjectMeta:{kube-controller-manager-crc.189827144aae4ce7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": read tcp 192.168.126.11:36310->192.168.126.11:10357: read: connection reset by peer Feb 27 16:24:03 crc kubenswrapper[4814]: body: Feb 27 16:24:03 crc kubenswrapper[4814]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:41.476056295 +0000 UTC m=+33.928681115,LastTimestamp:2026-02-27 16:23:41.476056295 +0000 UTC m=+33.928681115,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 16:24:03 crc kubenswrapper[4814]: > Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.463841 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189827144aaf118d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:36310->192.168.126.11:10357: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:41.476106637 +0000 UTC m=+33.928731467,LastTimestamp:2026-02-27 16:23:41.476106637 +0000 UTC m=+33.928731467,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.470034 4814 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189827144adbdac5 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:41.479041733 +0000 UTC m=+33.931666613,LastTimestamp:2026-02-27 16:23:41.479041733 +0000 UTC m=+33.931666613,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.474797 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.1898270ce5991459\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270ce5991459 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:09.715395673 +0000 UTC m=+2.168020523,LastTimestamp:2026-02-27 16:23:42.016221497 +0000 UTC m=+34.468846367,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.478924 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.1898270cf719a684\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270cf719a684 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.009034372 +0000 UTC m=+2.461659242,LastTimestamp:2026-02-27 16:23:42.275116753 +0000 UTC m=+34.727741623,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.485224 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.1898270cf7dc3d93\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270cf7dc3d93 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:10.021787027 +0000 UTC m=+2.474411857,LastTimestamp:2026-02-27 16:23:42.29955838 +0000 UTC m=+34.752183230,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.491330 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.1898270fe9792d96\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 16:24:03 crc kubenswrapper[4814]: &Event{ObjectMeta:{kube-controller-manager-crc.1898270fe9792d96 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 27 16:24:03 crc kubenswrapper[4814]: body: Feb 27 16:24:03 crc kubenswrapper[4814]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:22.665315734 +0000 UTC m=+15.117940624,LastTimestamp:2026-02-27 16:23:52.665762676 +0000 UTC m=+45.118387546,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 16:24:03 crc kubenswrapper[4814]: > Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.496901 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.1898270fe97a4228\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898270fe97a4228 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:22.665386536 +0000 UTC m=+15.118011406,LastTimestamp:2026-02-27 16:23:52.665821618 +0000 UTC m=+45.118446478,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:24:03 crc kubenswrapper[4814]: E0227 16:24:03.503337 4814 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.1898270fe9792d96\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 16:24:03 crc kubenswrapper[4814]: &Event{ObjectMeta:{kube-controller-manager-crc.1898270fe9792d96 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 27 16:24:03 crc kubenswrapper[4814]: body: Feb 27 16:24:03 crc kubenswrapper[4814]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:23:22.665315734 +0000 UTC m=+15.117940624,LastTimestamp:2026-02-27 16:24:02.665037142 +0000 UTC m=+55.117662012,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 16:24:03 crc kubenswrapper[4814]: > Feb 27 16:24:04 crc kubenswrapper[4814]: W0227 16:24:04.219597 4814 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:04 crc kubenswrapper[4814]: E0227 16:24:04.219668 4814 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Feb 27 16:24:04 crc kubenswrapper[4814]: I0227 16:24:04.233682 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:24:04 crc kubenswrapper[4814]: I0227 16:24:04.235494 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:04 crc kubenswrapper[4814]: I0227 16:24:04.235560 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:04 crc kubenswrapper[4814]: I0227 16:24:04.235580 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:04 crc kubenswrapper[4814]: I0227 16:24:04.235624 4814 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 16:24:04 crc kubenswrapper[4814]: E0227 16:24:04.245169 4814 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 16:24:04 crc kubenswrapper[4814]: E0227 16:24:04.245311 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 16:24:04 crc kubenswrapper[4814]: I0227 16:24:04.404931 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:04 crc kubenswrapper[4814]: I0227 16:24:04.507840 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 16:24:04 crc kubenswrapper[4814]: I0227 16:24:04.508144 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:24:04 crc kubenswrapper[4814]: I0227 16:24:04.510210 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:04 crc kubenswrapper[4814]: I0227 16:24:04.510273 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:04 crc kubenswrapper[4814]: I0227 16:24:04.510282 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:05 crc kubenswrapper[4814]: I0227 16:24:05.405161 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:06 crc kubenswrapper[4814]: I0227 16:24:06.404350 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:07 crc kubenswrapper[4814]: I0227 16:24:07.403806 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:07 crc kubenswrapper[4814]: I0227 16:24:07.487299 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:24:07 crc kubenswrapper[4814]: I0227 16:24:07.488682 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:07 crc kubenswrapper[4814]: I0227 16:24:07.488742 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:07 crc kubenswrapper[4814]: I0227 16:24:07.488757 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:07 crc kubenswrapper[4814]: I0227 16:24:07.489491 4814 scope.go:117] "RemoveContainer" containerID="fbc8df752c45f857b0470acc6fa6ebd84ca0eeb96a33a6fdb790bfd28df3f07c" Feb 27 16:24:07 crc kubenswrapper[4814]: E0227 16:24:07.489698 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 16:24:08 crc kubenswrapper[4814]: I0227 16:24:08.403007 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:08 crc kubenswrapper[4814]: E0227 16:24:08.576291 4814 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 16:24:09 crc kubenswrapper[4814]: I0227 16:24:09.404324 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:10 crc kubenswrapper[4814]: I0227 16:24:10.408564 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:11 crc kubenswrapper[4814]: I0227 16:24:11.245545 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:24:11 crc kubenswrapper[4814]: I0227 16:24:11.247525 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:11 crc kubenswrapper[4814]: I0227 16:24:11.247564 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:11 crc kubenswrapper[4814]: I0227 16:24:11.247574 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:11 crc kubenswrapper[4814]: I0227 16:24:11.247605 4814 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 16:24:11 crc kubenswrapper[4814]: E0227 16:24:11.252090 4814 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 16:24:11 crc kubenswrapper[4814]: E0227 16:24:11.252142 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 16:24:11 crc kubenswrapper[4814]: I0227 16:24:11.403480 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:11 crc kubenswrapper[4814]: I0227 16:24:11.668113 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:24:11 crc kubenswrapper[4814]: I0227 16:24:11.668331 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:24:11 crc kubenswrapper[4814]: I0227 16:24:11.669271 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:11 crc kubenswrapper[4814]: I0227 16:24:11.669290 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:11 crc kubenswrapper[4814]: I0227 16:24:11.669299 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:11 crc kubenswrapper[4814]: I0227 16:24:11.673082 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:24:11 crc kubenswrapper[4814]: I0227 16:24:11.774624 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:24:11 crc kubenswrapper[4814]: I0227 16:24:11.775873 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:11 crc kubenswrapper[4814]: I0227 16:24:11.775909 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:11 crc kubenswrapper[4814]: I0227 16:24:11.775919 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:12 crc kubenswrapper[4814]: I0227 16:24:12.402344 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:13 crc kubenswrapper[4814]: I0227 16:24:13.401496 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:14 crc kubenswrapper[4814]: I0227 16:24:14.402300 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:15 crc kubenswrapper[4814]: I0227 16:24:15.401402 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:16 crc kubenswrapper[4814]: I0227 16:24:16.402443 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:17 crc kubenswrapper[4814]: I0227 16:24:17.403617 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:18 crc kubenswrapper[4814]: I0227 16:24:18.252754 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:24:18 crc kubenswrapper[4814]: I0227 16:24:18.253803 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:18 crc kubenswrapper[4814]: I0227 16:24:18.253834 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:18 crc kubenswrapper[4814]: I0227 16:24:18.253845 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:18 crc kubenswrapper[4814]: I0227 16:24:18.253866 4814 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 16:24:18 crc kubenswrapper[4814]: E0227 16:24:18.259451 4814 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 16:24:18 crc kubenswrapper[4814]: E0227 16:24:18.259711 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 16:24:18 crc kubenswrapper[4814]: I0227 16:24:18.403912 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:18 crc kubenswrapper[4814]: I0227 16:24:18.487019 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:24:18 crc kubenswrapper[4814]: I0227 16:24:18.488611 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:18 crc kubenswrapper[4814]: I0227 16:24:18.488671 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:18 crc kubenswrapper[4814]: I0227 16:24:18.488692 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:18 crc kubenswrapper[4814]: I0227 16:24:18.489623 4814 scope.go:117] "RemoveContainer" containerID="fbc8df752c45f857b0470acc6fa6ebd84ca0eeb96a33a6fdb790bfd28df3f07c" Feb 27 16:24:18 crc kubenswrapper[4814]: E0227 16:24:18.576412 4814 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 16:24:19 crc kubenswrapper[4814]: I0227 16:24:19.406115 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:19 crc kubenswrapper[4814]: I0227 16:24:19.795639 4814 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 16:24:19 crc kubenswrapper[4814]: I0227 16:24:19.799486 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 16:24:19 crc kubenswrapper[4814]: I0227 16:24:19.800043 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 27 16:24:19 crc kubenswrapper[4814]: I0227 16:24:19.802330 4814 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1" exitCode=255 Feb 27 16:24:19 crc kubenswrapper[4814]: I0227 16:24:19.802366 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1"} Feb 27 16:24:19 crc kubenswrapper[4814]: I0227 16:24:19.802398 4814 scope.go:117] "RemoveContainer" containerID="fbc8df752c45f857b0470acc6fa6ebd84ca0eeb96a33a6fdb790bfd28df3f07c" Feb 27 16:24:19 crc kubenswrapper[4814]: I0227 16:24:19.802587 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:24:19 crc kubenswrapper[4814]: I0227 16:24:19.803868 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:19 crc kubenswrapper[4814]: I0227 16:24:19.803895 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:19 crc kubenswrapper[4814]: I0227 16:24:19.803903 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:19 crc kubenswrapper[4814]: I0227 16:24:19.804380 4814 scope.go:117] "RemoveContainer" containerID="bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1" Feb 27 16:24:19 crc kubenswrapper[4814]: E0227 16:24:19.804533 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 16:24:19 crc kubenswrapper[4814]: I0227 16:24:19.817674 4814 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 27 16:24:20 crc kubenswrapper[4814]: I0227 16:24:20.280215 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:24:20 crc kubenswrapper[4814]: I0227 16:24:20.402048 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:20 crc kubenswrapper[4814]: I0227 16:24:20.807858 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 16:24:20 crc kubenswrapper[4814]: I0227 16:24:20.811438 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:24:20 crc kubenswrapper[4814]: I0227 16:24:20.812752 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:20 crc kubenswrapper[4814]: I0227 16:24:20.812828 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:20 crc kubenswrapper[4814]: I0227 16:24:20.812856 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:20 crc kubenswrapper[4814]: I0227 16:24:20.813671 4814 scope.go:117] "RemoveContainer" containerID="bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1" Feb 27 16:24:20 crc kubenswrapper[4814]: E0227 16:24:20.813941 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 16:24:21 crc kubenswrapper[4814]: I0227 16:24:21.404190 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:22 crc kubenswrapper[4814]: I0227 16:24:22.404831 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:22 crc kubenswrapper[4814]: I0227 16:24:22.486835 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:24:22 crc kubenswrapper[4814]: I0227 16:24:22.488575 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:22 crc kubenswrapper[4814]: I0227 16:24:22.488659 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:22 crc kubenswrapper[4814]: I0227 16:24:22.488684 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:23 crc kubenswrapper[4814]: I0227 16:24:23.404104 4814 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 16:24:23 crc kubenswrapper[4814]: I0227 16:24:23.416312 4814 csr.go:261] certificate signing request csr-fvwzb is approved, waiting to be issued Feb 27 16:24:23 crc kubenswrapper[4814]: I0227 16:24:23.428605 4814 csr.go:257] certificate signing request csr-fvwzb is issued Feb 27 16:24:23 crc kubenswrapper[4814]: I0227 16:24:23.439786 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:24:23 crc kubenswrapper[4814]: I0227 16:24:23.440033 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:24:23 crc kubenswrapper[4814]: I0227 16:24:23.441579 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:23 crc kubenswrapper[4814]: I0227 16:24:23.441629 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:23 crc kubenswrapper[4814]: I0227 16:24:23.441646 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:23 crc kubenswrapper[4814]: I0227 16:24:23.442534 4814 scope.go:117] "RemoveContainer" containerID="bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1" Feb 27 16:24:23 crc kubenswrapper[4814]: E0227 16:24:23.442821 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 16:24:23 crc kubenswrapper[4814]: I0227 16:24:23.476422 4814 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 27 16:24:24 crc kubenswrapper[4814]: I0227 16:24:24.141955 4814 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 27 16:24:24 crc kubenswrapper[4814]: I0227 16:24:24.429969 4814 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-16 19:57:15.125788203 +0000 UTC Feb 27 16:24:24 crc kubenswrapper[4814]: I0227 16:24:24.430036 4814 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7011h32m50.695757781s for next certificate rotation Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.260247 4814 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.261923 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.261974 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.261991 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.262245 4814 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.273223 4814 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.273779 4814 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 27 16:24:25 crc kubenswrapper[4814]: E0227 16:24:25.273926 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.279398 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.279461 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.279479 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.279502 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.279519 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:25Z","lastTransitionTime":"2026-02-27T16:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:25 crc kubenswrapper[4814]: E0227 16:24:25.299942 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.311012 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.311074 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.311099 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.311132 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.311156 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:25Z","lastTransitionTime":"2026-02-27T16:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:25 crc kubenswrapper[4814]: E0227 16:24:25.327650 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.336003 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.336045 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.336062 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.336099 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.336137 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:25Z","lastTransitionTime":"2026-02-27T16:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:25 crc kubenswrapper[4814]: E0227 16:24:25.352769 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.364404 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.364480 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.364513 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.364542 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:25 crc kubenswrapper[4814]: I0227 16:24:25.364562 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:25Z","lastTransitionTime":"2026-02-27T16:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:25 crc kubenswrapper[4814]: E0227 16:24:25.378632 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 16:24:25 crc kubenswrapper[4814]: E0227 16:24:25.378910 4814 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 16:24:25 crc kubenswrapper[4814]: E0227 16:24:25.378953 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:25 crc kubenswrapper[4814]: E0227 16:24:25.480046 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:25 crc kubenswrapper[4814]: E0227 16:24:25.580317 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:25 crc kubenswrapper[4814]: E0227 16:24:25.680673 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:25 crc kubenswrapper[4814]: E0227 16:24:25.781122 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:25 crc kubenswrapper[4814]: E0227 16:24:25.882179 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:25 crc kubenswrapper[4814]: E0227 16:24:25.982751 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:26 crc kubenswrapper[4814]: E0227 16:24:26.083515 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:26 crc kubenswrapper[4814]: E0227 16:24:26.184504 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:26 crc kubenswrapper[4814]: E0227 16:24:26.284724 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:26 crc kubenswrapper[4814]: E0227 16:24:26.385620 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:26 crc kubenswrapper[4814]: E0227 16:24:26.486398 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:26 crc kubenswrapper[4814]: E0227 16:24:26.587628 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:26 crc kubenswrapper[4814]: E0227 16:24:26.688672 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:26 crc kubenswrapper[4814]: E0227 16:24:26.789773 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:26 crc kubenswrapper[4814]: E0227 16:24:26.890241 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:26 crc kubenswrapper[4814]: E0227 16:24:26.991304 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:27 crc kubenswrapper[4814]: E0227 16:24:27.091957 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:27 crc kubenswrapper[4814]: I0227 16:24:27.172232 4814 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 27 16:24:27 crc kubenswrapper[4814]: E0227 16:24:27.192433 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:27 crc kubenswrapper[4814]: E0227 16:24:27.293591 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:27 crc kubenswrapper[4814]: E0227 16:24:27.394375 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:27 crc kubenswrapper[4814]: E0227 16:24:27.495361 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:27 crc kubenswrapper[4814]: E0227 16:24:27.596155 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:27 crc kubenswrapper[4814]: E0227 16:24:27.696566 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:27 crc kubenswrapper[4814]: E0227 16:24:27.797599 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:27 crc kubenswrapper[4814]: E0227 16:24:27.898592 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:27 crc kubenswrapper[4814]: E0227 16:24:27.999500 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:28 crc kubenswrapper[4814]: E0227 16:24:28.100399 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:28 crc kubenswrapper[4814]: E0227 16:24:28.200720 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:28 crc kubenswrapper[4814]: E0227 16:24:28.301691 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:28 crc kubenswrapper[4814]: E0227 16:24:28.402010 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:28 crc kubenswrapper[4814]: I0227 16:24:28.466498 4814 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 27 16:24:28 crc kubenswrapper[4814]: E0227 16:24:28.502919 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:28 crc kubenswrapper[4814]: E0227 16:24:28.577730 4814 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 16:24:28 crc kubenswrapper[4814]: E0227 16:24:28.603457 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:28 crc kubenswrapper[4814]: E0227 16:24:28.703695 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:28 crc kubenswrapper[4814]: E0227 16:24:28.804826 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:28 crc kubenswrapper[4814]: E0227 16:24:28.905497 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:29 crc kubenswrapper[4814]: E0227 16:24:29.005679 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:29 crc kubenswrapper[4814]: E0227 16:24:29.106502 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:29 crc kubenswrapper[4814]: E0227 16:24:29.207281 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:29 crc kubenswrapper[4814]: E0227 16:24:29.308373 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:29 crc kubenswrapper[4814]: E0227 16:24:29.408834 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:29 crc kubenswrapper[4814]: E0227 16:24:29.509230 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:29 crc kubenswrapper[4814]: E0227 16:24:29.609834 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:29 crc kubenswrapper[4814]: E0227 16:24:29.711108 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:29 crc kubenswrapper[4814]: E0227 16:24:29.811740 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:29 crc kubenswrapper[4814]: E0227 16:24:29.912288 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:30 crc kubenswrapper[4814]: E0227 16:24:30.013174 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:30 crc kubenswrapper[4814]: E0227 16:24:30.113387 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:30 crc kubenswrapper[4814]: E0227 16:24:30.214634 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:30 crc kubenswrapper[4814]: E0227 16:24:30.315688 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:30 crc kubenswrapper[4814]: E0227 16:24:30.416102 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:30 crc kubenswrapper[4814]: E0227 16:24:30.516499 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:30 crc kubenswrapper[4814]: E0227 16:24:30.617345 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:30 crc kubenswrapper[4814]: E0227 16:24:30.717540 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:30 crc kubenswrapper[4814]: E0227 16:24:30.818304 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:30 crc kubenswrapper[4814]: E0227 16:24:30.918543 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:31 crc kubenswrapper[4814]: E0227 16:24:31.019166 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:31 crc kubenswrapper[4814]: E0227 16:24:31.119987 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:31 crc kubenswrapper[4814]: E0227 16:24:31.220194 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:31 crc kubenswrapper[4814]: E0227 16:24:31.321311 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:31 crc kubenswrapper[4814]: E0227 16:24:31.422232 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:31 crc kubenswrapper[4814]: E0227 16:24:31.522361 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:31 crc kubenswrapper[4814]: E0227 16:24:31.623486 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:31 crc kubenswrapper[4814]: E0227 16:24:31.723632 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:31 crc kubenswrapper[4814]: E0227 16:24:31.824398 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:31 crc kubenswrapper[4814]: E0227 16:24:31.924563 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:32 crc kubenswrapper[4814]: E0227 16:24:32.024755 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:32 crc kubenswrapper[4814]: E0227 16:24:32.125395 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:32 crc kubenswrapper[4814]: E0227 16:24:32.225592 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:32 crc kubenswrapper[4814]: E0227 16:24:32.326580 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:32 crc kubenswrapper[4814]: E0227 16:24:32.426991 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:32 crc kubenswrapper[4814]: E0227 16:24:32.527612 4814 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.607813 4814 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.629920 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.629971 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.629984 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.630002 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.630014 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:32Z","lastTransitionTime":"2026-02-27T16:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.732830 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.732873 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.732882 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.732895 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.732909 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:32Z","lastTransitionTime":"2026-02-27T16:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.835820 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.835869 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.835886 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.835910 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.835929 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:32Z","lastTransitionTime":"2026-02-27T16:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.939208 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.939310 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.939330 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.939354 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:32 crc kubenswrapper[4814]: I0227 16:24:32.939371 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:32Z","lastTransitionTime":"2026-02-27T16:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.042794 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.042857 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.042890 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.042919 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.042938 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:33Z","lastTransitionTime":"2026-02-27T16:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.146068 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.146137 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.146162 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.146190 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.146210 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:33Z","lastTransitionTime":"2026-02-27T16:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.249205 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.249274 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.249286 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.249305 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.249318 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:33Z","lastTransitionTime":"2026-02-27T16:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.352370 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.352435 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.352458 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.352489 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.352520 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:33Z","lastTransitionTime":"2026-02-27T16:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.415132 4814 apiserver.go:52] "Watching apiserver" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.422106 4814 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.422642 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.423317 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.423532 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.423711 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 16:24:33 crc kubenswrapper[4814]: E0227 16:24:33.424047 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.424091 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.424191 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:33 crc kubenswrapper[4814]: E0227 16:24:33.424779 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.426596 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.426746 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.429427 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.429557 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.429717 4814 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.429720 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.429844 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.429906 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.429852 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.429851 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.430500 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:33 crc kubenswrapper[4814]: E0227 16:24:33.430636 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.445400 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.445480 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.445533 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.445576 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.445622 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.445677 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.445721 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.445765 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.445813 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.445864 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.445915 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446053 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446112 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446162 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446207 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446292 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446345 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446425 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446478 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446531 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446516 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446580 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446636 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446674 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446739 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446777 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446809 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446842 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446877 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446923 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.446977 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447028 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447073 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447108 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447140 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447141 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447174 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447213 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447250 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447297 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447318 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447438 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447509 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447578 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447631 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447685 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447854 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447896 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447913 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.447973 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.448030 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.448086 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.448145 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.448201 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.448294 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.448347 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.448402 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.448456 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.448506 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.448443 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.448546 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.448531 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.449043 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.449059 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.449174 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.449206 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.449683 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.449884 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.449898 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.450047 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.450301 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.450509 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.450569 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.451127 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.451334 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.452230 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.452281 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.452352 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.452783 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.453341 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.453417 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.453738 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.453909 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.453928 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.454542 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.455093 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.455192 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.455757 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.456208 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.456642 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.456670 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.457467 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.457872 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.458013 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.458217 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.458298 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.459518 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.459970 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.460450 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.460612 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.448563 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.460713 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.460781 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.460822 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.460820 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.460858 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.460935 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.460972 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461008 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461047 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461083 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461121 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461157 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461212 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461248 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461310 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461350 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461426 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461462 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461500 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461533 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461566 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461609 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461647 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461688 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461722 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461757 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461791 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461824 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461857 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461912 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461945 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461986 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462023 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462078 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462113 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462151 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462188 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462225 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462285 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462325 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462361 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462396 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462448 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462485 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462521 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462559 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462593 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462632 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462666 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462699 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462734 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462769 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462807 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462842 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462878 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462936 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462971 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463009 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463065 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463106 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463142 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463182 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463216 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463275 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463313 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463348 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463383 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463452 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463490 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463526 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463568 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463626 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463664 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463700 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463735 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463776 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463880 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463919 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463957 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463994 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.464035 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.464086 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.464127 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.464166 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465005 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465054 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465089 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465126 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465167 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465204 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465241 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465306 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465343 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465405 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465447 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465482 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465520 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465559 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465595 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465634 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465674 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465811 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465857 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465896 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465933 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465972 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466009 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466045 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466080 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466115 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466153 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466191 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466228 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466321 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466361 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466399 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466442 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466482 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466518 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466556 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466684 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466723 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466762 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466799 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466837 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466877 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466916 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466963 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467002 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467045 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467083 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467121 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467157 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467194 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467230 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467295 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467334 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467374 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467414 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467454 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467533 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467588 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467631 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468004 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468055 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468101 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468143 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468186 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468227 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468293 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461558 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468336 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468383 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468423 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468461 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468692 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468720 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468744 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468767 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468790 4814 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469353 4814 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469405 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469429 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469450 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469474 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469497 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469519 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469542 4814 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469563 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469585 4814 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469609 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469632 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469654 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469674 4814 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469696 4814 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469717 4814 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469738 4814 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469759 4814 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469780 4814 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469802 4814 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469825 4814 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469848 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469869 4814 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469893 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469917 4814 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469941 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469962 4814 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469985 4814 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.470006 4814 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.470028 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.470052 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.470074 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.470096 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.470120 4814 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.470143 4814 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.470163 4814 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.470184 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.470207 4814 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.470228 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.470250 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.470299 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.470319 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.488003 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.488059 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.488076 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.488124 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.488147 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:33Z","lastTransitionTime":"2026-02-27T16:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.488433 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.497506 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.461917 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462424 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463381 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463553 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.463695 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.462540 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.464873 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465602 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.465621 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466099 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466308 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.466899 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467474 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.467793 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468293 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468502 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.468951 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469541 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.469859 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.470525 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.471037 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.471558 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.471652 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.471760 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.471777 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.472007 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.472137 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.472329 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.473041 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.473687 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.473714 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.473980 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.474465 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.474747 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.475382 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.501820 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.475494 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.501887 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.475519 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.476219 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.476385 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.476870 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.477809 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.477923 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.478023 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.478102 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.477741 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.478478 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.478118 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.478643 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.478965 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.478973 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.479042 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.479088 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.479197 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.479743 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.480704 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.481876 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.481928 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.483115 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.484000 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.484807 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.484828 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.485093 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.485354 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.485769 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.485839 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.485866 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.485863 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.485889 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.485897 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.486202 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.486449 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.486381 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.487736 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.487753 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.487770 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.488736 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.488969 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.489308 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.489499 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.489534 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.489571 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.489896 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.490221 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.490832 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.490952 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.491190 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.491834 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.492311 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.492364 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: E0227 16:24:33.493739 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:24:33.993700612 +0000 UTC m=+86.446325482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.497774 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.497889 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.497983 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.498046 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: E0227 16:24:33.498421 4814 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.498461 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.498966 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.499938 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.499961 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.500016 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.500026 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.500355 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.500542 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.500647 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.501081 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.501944 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: E0227 16:24:33.502015 4814 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.503062 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.503653 4814 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 27 16:24:33 crc kubenswrapper[4814]: E0227 16:24:33.503702 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:34.003648546 +0000 UTC m=+86.456273416 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:24:33 crc kubenswrapper[4814]: E0227 16:24:33.503991 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:34.003975887 +0000 UTC m=+86.456600747 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.504096 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.502078 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.505796 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.510122 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.510382 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.512702 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.516391 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.516707 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.519016 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.520671 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.527629 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.528641 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.528803 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.529140 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.533635 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 16:24:33 crc kubenswrapper[4814]: E0227 16:24:33.535605 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:24:33 crc kubenswrapper[4814]: E0227 16:24:33.535649 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:24:33 crc kubenswrapper[4814]: E0227 16:24:33.535705 4814 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:33 crc kubenswrapper[4814]: E0227 16:24:33.535818 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:34.035787106 +0000 UTC m=+86.488411966 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.535937 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.537775 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.538785 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.539038 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.539447 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.539873 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.539954 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.540903 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.540919 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.541050 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.541531 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.541864 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.543576 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.545503 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.545705 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.545743 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.546448 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.546828 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.547470 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.550788 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.550834 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.550952 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.551212 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.551373 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.551407 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: E0227 16:24:33.551470 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:24:33 crc kubenswrapper[4814]: E0227 16:24:33.551511 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:24:33 crc kubenswrapper[4814]: E0227 16:24:33.551538 4814 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:33 crc kubenswrapper[4814]: E0227 16:24:33.551631 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:34.051598773 +0000 UTC m=+86.504223783 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.551715 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.551942 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.557726 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.558748 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.559525 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.571449 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.571526 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.571728 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.571760 4814 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.571786 4814 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.571809 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.571832 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.571855 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.571878 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.571903 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.571927 4814 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.571948 4814 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.571971 4814 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.571997 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.572018 4814 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.572040 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.572063 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.572085 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.572108 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.572132 4814 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.572358 4814 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.572391 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.572418 4814 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.572443 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573621 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573652 4814 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573675 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573701 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573726 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573749 4814 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573773 4814 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573797 4814 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573148 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573821 4814 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573846 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573871 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573895 4814 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573920 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573942 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573965 4814 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573990 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574014 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574037 4814 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574060 4814 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574084 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574108 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574130 4814 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574154 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574178 4814 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574200 4814 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574226 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574249 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574310 4814 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574332 4814 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574357 4814 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574381 4814 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574407 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574430 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574457 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574479 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.573564 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574849 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574843 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.574882 4814 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575107 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575130 4814 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575145 4814 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575162 4814 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575177 4814 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575192 4814 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575206 4814 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575221 4814 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575235 4814 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575250 4814 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575296 4814 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575310 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575324 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575338 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575353 4814 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575367 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575382 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575398 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575414 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575429 4814 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575443 4814 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575458 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575472 4814 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575488 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575501 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575514 4814 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575527 4814 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575540 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575555 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575569 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575584 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575600 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575613 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575629 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575643 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575657 4814 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575672 4814 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575685 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575701 4814 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575715 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575728 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575743 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575757 4814 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575770 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575785 4814 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575799 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575813 4814 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575830 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575844 4814 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575858 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575873 4814 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575887 4814 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575901 4814 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575916 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575931 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575945 4814 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575958 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575971 4814 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.575986 4814 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576001 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576015 4814 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576030 4814 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576044 4814 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576059 4814 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576136 4814 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576152 4814 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576166 4814 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576181 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576194 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576208 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576221 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576234 4814 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576249 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576308 4814 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576322 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576335 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576349 4814 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576340 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.576362 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.577215 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.577233 4814 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.577247 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.577277 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.577295 4814 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.577308 4814 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.579225 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.584414 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.586664 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.587483 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.591689 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.591728 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.591739 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.591759 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.591779 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:33Z","lastTransitionTime":"2026-02-27T16:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.602502 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.619115 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.678186 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.678229 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.678273 4814 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.678292 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.694524 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.694602 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.694616 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.694642 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.694663 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:33Z","lastTransitionTime":"2026-02-27T16:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.756805 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.779299 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.797645 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.797717 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.797735 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.797767 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.797787 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:33Z","lastTransitionTime":"2026-02-27T16:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:33 crc kubenswrapper[4814]: W0227 16:24:33.798230 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-7ccc25e9e1e890ac62deb4af2ac2cd367040a653102a5680073fd224ceb2936c WatchSource:0}: Error finding container 7ccc25e9e1e890ac62deb4af2ac2cd367040a653102a5680073fd224ceb2936c: Status 404 returned error can't find the container with id 7ccc25e9e1e890ac62deb4af2ac2cd367040a653102a5680073fd224ceb2936c Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.799967 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 16:24:33 crc kubenswrapper[4814]: W0227 16:24:33.825627 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-e00188f7a171995e0ae352b2b0c43533a8e6cd972016310919ada8ecf230e59f WatchSource:0}: Error finding container e00188f7a171995e0ae352b2b0c43533a8e6cd972016310919ada8ecf230e59f: Status 404 returned error can't find the container with id e00188f7a171995e0ae352b2b0c43533a8e6cd972016310919ada8ecf230e59f Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.851669 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b1ff95e2af99cf37cf61be273532e410760111dc28e66ef255936726764ed94c"} Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.855116 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e00188f7a171995e0ae352b2b0c43533a8e6cd972016310919ada8ecf230e59f"} Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.856500 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7ccc25e9e1e890ac62deb4af2ac2cd367040a653102a5680073fd224ceb2936c"} Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.901564 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.901652 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.901670 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.901720 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:33 crc kubenswrapper[4814]: I0227 16:24:33.901740 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:33Z","lastTransitionTime":"2026-02-27T16:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.005275 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.005323 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.005348 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.005369 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.005381 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:34Z","lastTransitionTime":"2026-02-27T16:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.082992 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.083153 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:34 crc kubenswrapper[4814]: E0227 16:24:34.083197 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:24:35.083169914 +0000 UTC m=+87.535794764 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.083231 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.083294 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.083331 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:34 crc kubenswrapper[4814]: E0227 16:24:34.083383 4814 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:24:34 crc kubenswrapper[4814]: E0227 16:24:34.083453 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:24:34 crc kubenswrapper[4814]: E0227 16:24:34.083471 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:24:34 crc kubenswrapper[4814]: E0227 16:24:34.083468 4814 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:24:34 crc kubenswrapper[4814]: E0227 16:24:34.083490 4814 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:34 crc kubenswrapper[4814]: E0227 16:24:34.083473 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:35.083438254 +0000 UTC m=+87.536063094 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:24:34 crc kubenswrapper[4814]: E0227 16:24:34.083491 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:24:34 crc kubenswrapper[4814]: E0227 16:24:34.083706 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:24:34 crc kubenswrapper[4814]: E0227 16:24:34.083725 4814 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:34 crc kubenswrapper[4814]: E0227 16:24:34.083674 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:35.083638451 +0000 UTC m=+87.536263301 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:24:34 crc kubenswrapper[4814]: E0227 16:24:34.083768 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:35.083759245 +0000 UTC m=+87.536384085 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:34 crc kubenswrapper[4814]: E0227 16:24:34.083784 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:35.083776015 +0000 UTC m=+87.536400855 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.108578 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.108617 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.108628 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.108646 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.108657 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:34Z","lastTransitionTime":"2026-02-27T16:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.212079 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.212123 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.212135 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.212155 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.212168 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:34Z","lastTransitionTime":"2026-02-27T16:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.315124 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.315192 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.315214 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.315303 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.315332 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:34Z","lastTransitionTime":"2026-02-27T16:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.417455 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.417502 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.417514 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.417530 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.417542 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:34Z","lastTransitionTime":"2026-02-27T16:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.490672 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.491187 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.492589 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.493334 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.494506 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.495326 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.495917 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.496833 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.497494 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.498400 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.498896 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.499913 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.500425 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.500962 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.502152 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.502692 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.503619 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.503984 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.504548 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.505651 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.506232 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.507411 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.508418 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.509571 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.509969 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.510582 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.511666 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.512129 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.513182 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.513705 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.514568 4814 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.514670 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.516352 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.517235 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.517684 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.519375 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.519951 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.519990 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.519998 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.520014 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.520024 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:34Z","lastTransitionTime":"2026-02-27T16:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.520114 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.521314 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.522147 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.523278 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.523728 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.524961 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.525714 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.526721 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.527283 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.528266 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.528830 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.530926 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.531588 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.532615 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.533173 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.534335 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.536129 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.536956 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.622288 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.622328 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.622339 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.622355 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.622368 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:34Z","lastTransitionTime":"2026-02-27T16:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.724522 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.724551 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.724561 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.724575 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.724584 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:34Z","lastTransitionTime":"2026-02-27T16:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.826468 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.826521 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.826532 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.826549 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.826557 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:34Z","lastTransitionTime":"2026-02-27T16:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.860459 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199"} Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.861312 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d"} Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.863315 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd"} Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.885189 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:34Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.902544 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:34Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.919187 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:34Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.929146 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.929220 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.929237 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.929301 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.929318 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:34Z","lastTransitionTime":"2026-02-27T16:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.936018 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:34Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.955179 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:34Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.970437 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:34Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.983682 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:34Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:34 crc kubenswrapper[4814]: I0227 16:24:34.996326 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:34Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.010448 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:35Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.024839 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:35Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.031992 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.032031 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.032042 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.032059 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.032070 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:35Z","lastTransitionTime":"2026-02-27T16:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.042412 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:35Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.057172 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:35Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.093777 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.093888 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.093917 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.093941 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.093970 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.094110 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.094134 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.094149 4814 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.094209 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:37.094191455 +0000 UTC m=+89.546816285 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.094638 4814 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.094676 4814 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.094667 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:24:37.094636371 +0000 UTC m=+89.547261201 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.094852 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:37.094824467 +0000 UTC m=+89.547449327 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.094851 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.094955 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.094875 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:37.094862448 +0000 UTC m=+89.547487308 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.094981 4814 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.095119 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:37.095086666 +0000 UTC m=+89.547711526 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.134700 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.134973 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.135057 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.135141 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.135212 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:35Z","lastTransitionTime":"2026-02-27T16:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.237959 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.237992 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.238003 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.238019 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.238031 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:35Z","lastTransitionTime":"2026-02-27T16:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.340649 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.340695 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.340711 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.340734 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.340756 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:35Z","lastTransitionTime":"2026-02-27T16:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.397916 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.398327 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.398478 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.398617 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.398737 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:35Z","lastTransitionTime":"2026-02-27T16:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.419598 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:35Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.428354 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.428409 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.428420 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.428440 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.428456 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:35Z","lastTransitionTime":"2026-02-27T16:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.447075 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:35Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.452319 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.452388 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.452405 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.452424 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.452447 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:35Z","lastTransitionTime":"2026-02-27T16:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.469654 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:35Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.473775 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.473809 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.473817 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.473831 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.473868 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:35Z","lastTransitionTime":"2026-02-27T16:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.487062 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.487034 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:35Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.487139 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.487080 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.487231 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.487335 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.487515 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.498999 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.499044 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.499058 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.499079 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.499094 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:35Z","lastTransitionTime":"2026-02-27T16:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.517563 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:35Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:35 crc kubenswrapper[4814]: E0227 16:24:35.517793 4814 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.520280 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.520334 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.520352 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.520378 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.520397 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:35Z","lastTransitionTime":"2026-02-27T16:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.622763 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.622805 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.622817 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.622832 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.622842 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:35Z","lastTransitionTime":"2026-02-27T16:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.724924 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.724971 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.724988 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.725008 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.725023 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:35Z","lastTransitionTime":"2026-02-27T16:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.826834 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.826859 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.826866 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.826878 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.826886 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:35Z","lastTransitionTime":"2026-02-27T16:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.929488 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.929532 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.929544 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.929558 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:35 crc kubenswrapper[4814]: I0227 16:24:35.929569 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:35Z","lastTransitionTime":"2026-02-27T16:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.031549 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.031627 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.031653 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.031686 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.031707 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:36Z","lastTransitionTime":"2026-02-27T16:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.134354 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.134389 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.134403 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.134420 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.134430 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:36Z","lastTransitionTime":"2026-02-27T16:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.236836 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.236880 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.236891 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.236912 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.236925 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:36Z","lastTransitionTime":"2026-02-27T16:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.339425 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.339462 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.339475 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.339495 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.339512 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:36Z","lastTransitionTime":"2026-02-27T16:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.442129 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.442207 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.442232 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.442282 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.442303 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:36Z","lastTransitionTime":"2026-02-27T16:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.545154 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.545230 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.545248 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.545308 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.545326 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:36Z","lastTransitionTime":"2026-02-27T16:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.648894 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.649003 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.649024 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.649055 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.649080 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:36Z","lastTransitionTime":"2026-02-27T16:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.752155 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.752221 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.752235 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.752275 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.752293 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:36Z","lastTransitionTime":"2026-02-27T16:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.792618 4814 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.853933 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.853959 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.853967 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.853979 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.853987 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:36Z","lastTransitionTime":"2026-02-27T16:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.868508 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6"} Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.892520 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:36Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.911406 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:36Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.929306 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:36Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.941229 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:36Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.955298 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:36Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.956271 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.956398 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.956474 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.956548 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.956636 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:36Z","lastTransitionTime":"2026-02-27T16:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:36 crc kubenswrapper[4814]: I0227 16:24:36.969952 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:36Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.059503 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.059590 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.059613 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.059645 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.059668 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:37Z","lastTransitionTime":"2026-02-27T16:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.116449 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.116554 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.116598 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:37 crc kubenswrapper[4814]: E0227 16:24:37.116631 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:24:41.116599149 +0000 UTC m=+93.569223989 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:24:37 crc kubenswrapper[4814]: E0227 16:24:37.116693 4814 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.116702 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:37 crc kubenswrapper[4814]: E0227 16:24:37.116791 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:24:37 crc kubenswrapper[4814]: E0227 16:24:37.116817 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:41.116735544 +0000 UTC m=+93.569360414 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:24:37 crc kubenswrapper[4814]: E0227 16:24:37.116827 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:24:37 crc kubenswrapper[4814]: E0227 16:24:37.116846 4814 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:37 crc kubenswrapper[4814]: E0227 16:24:37.116848 4814 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.116873 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:37 crc kubenswrapper[4814]: E0227 16:24:37.116897 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:41.11688781 +0000 UTC m=+93.569512650 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:24:37 crc kubenswrapper[4814]: E0227 16:24:37.116916 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:41.11690739 +0000 UTC m=+93.569532230 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:37 crc kubenswrapper[4814]: E0227 16:24:37.116985 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:24:37 crc kubenswrapper[4814]: E0227 16:24:37.117005 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:24:37 crc kubenswrapper[4814]: E0227 16:24:37.117019 4814 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:37 crc kubenswrapper[4814]: E0227 16:24:37.117060 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:41.117043945 +0000 UTC m=+93.569668885 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.162752 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.162807 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.162824 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.162849 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.162869 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:37Z","lastTransitionTime":"2026-02-27T16:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.265741 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.265785 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.265797 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.265815 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.265829 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:37Z","lastTransitionTime":"2026-02-27T16:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.368623 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.368668 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.368677 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.368691 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.368700 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:37Z","lastTransitionTime":"2026-02-27T16:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.470933 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.471006 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.471031 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.471060 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.471085 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:37Z","lastTransitionTime":"2026-02-27T16:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.487126 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.487170 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:37 crc kubenswrapper[4814]: E0227 16:24:37.487431 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.487170 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:37 crc kubenswrapper[4814]: E0227 16:24:37.487553 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:24:37 crc kubenswrapper[4814]: E0227 16:24:37.487665 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.573717 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.573763 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.573775 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.573792 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.573804 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:37Z","lastTransitionTime":"2026-02-27T16:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.676433 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.676472 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.676481 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.676495 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.676506 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:37Z","lastTransitionTime":"2026-02-27T16:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.779177 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.779248 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.779303 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.779334 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.779356 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:37Z","lastTransitionTime":"2026-02-27T16:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.881857 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.881895 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.881904 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.881914 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.881924 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:37Z","lastTransitionTime":"2026-02-27T16:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.984602 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.984670 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.984686 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.984714 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:37 crc kubenswrapper[4814]: I0227 16:24:37.984733 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:37Z","lastTransitionTime":"2026-02-27T16:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.088397 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.089135 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.089235 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.089347 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.089429 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:38Z","lastTransitionTime":"2026-02-27T16:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.192291 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.192614 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.192753 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.192970 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.193153 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:38Z","lastTransitionTime":"2026-02-27T16:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.295942 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.296215 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.296332 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.296431 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.296515 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:38Z","lastTransitionTime":"2026-02-27T16:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.399487 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.399548 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.399565 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.399589 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.399607 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:38Z","lastTransitionTime":"2026-02-27T16:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.501892 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.501934 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.501946 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.501961 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.501972 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:38Z","lastTransitionTime":"2026-02-27T16:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.502743 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.505312 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.506404 4814 scope.go:117] "RemoveContainer" containerID="bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1" Feb 27 16:24:38 crc kubenswrapper[4814]: E0227 16:24:38.506732 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.519797 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.531050 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.541490 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.553885 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.565107 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.603433 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.603491 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.603510 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.603534 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.603554 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:38Z","lastTransitionTime":"2026-02-27T16:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.707793 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.707949 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.707969 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.707995 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.708016 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:38Z","lastTransitionTime":"2026-02-27T16:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.810843 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.810908 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.810927 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.810953 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.810972 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:38Z","lastTransitionTime":"2026-02-27T16:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.875485 4814 scope.go:117] "RemoveContainer" containerID="bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1" Feb 27 16:24:38 crc kubenswrapper[4814]: E0227 16:24:38.875850 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.913872 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.913950 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.913963 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.913996 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:38 crc kubenswrapper[4814]: I0227 16:24:38.914017 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:38Z","lastTransitionTime":"2026-02-27T16:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.017336 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.017401 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.017420 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.017443 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.017461 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:39Z","lastTransitionTime":"2026-02-27T16:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.120697 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.120785 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.120809 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.120841 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.120865 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:39Z","lastTransitionTime":"2026-02-27T16:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.223400 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.223475 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.223497 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.223527 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.223553 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:39Z","lastTransitionTime":"2026-02-27T16:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.325896 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.325994 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.326013 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.326044 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.326066 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:39Z","lastTransitionTime":"2026-02-27T16:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.427968 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.428027 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.428043 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.428069 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.428086 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:39Z","lastTransitionTime":"2026-02-27T16:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.488160 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:39 crc kubenswrapper[4814]: E0227 16:24:39.488390 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.488213 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:39 crc kubenswrapper[4814]: E0227 16:24:39.488474 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.488168 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:39 crc kubenswrapper[4814]: E0227 16:24:39.488533 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.530962 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.531191 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.531318 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.531411 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.531492 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:39Z","lastTransitionTime":"2026-02-27T16:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.634140 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.634213 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.634243 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.634285 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.634299 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:39Z","lastTransitionTime":"2026-02-27T16:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.736968 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.737023 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.737041 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.737060 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.737072 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:39Z","lastTransitionTime":"2026-02-27T16:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.839597 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.839661 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.839679 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.839704 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.839722 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:39Z","lastTransitionTime":"2026-02-27T16:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.942761 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.942821 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.942839 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.942863 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:39 crc kubenswrapper[4814]: I0227 16:24:39.942883 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:39Z","lastTransitionTime":"2026-02-27T16:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.045713 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.045772 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.045790 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.045816 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.045832 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:40Z","lastTransitionTime":"2026-02-27T16:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.149141 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.149189 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.149206 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.149230 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.149248 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:40Z","lastTransitionTime":"2026-02-27T16:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.252750 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.252932 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.253000 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.253035 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.253091 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:40Z","lastTransitionTime":"2026-02-27T16:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.357107 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.357306 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.357334 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.357396 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.357419 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:40Z","lastTransitionTime":"2026-02-27T16:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.460245 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.460337 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.460347 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.460364 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.460374 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:40Z","lastTransitionTime":"2026-02-27T16:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.563105 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.563173 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.563192 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.563220 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.563238 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:40Z","lastTransitionTime":"2026-02-27T16:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.666579 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.666624 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.666637 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.666655 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.666668 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:40Z","lastTransitionTime":"2026-02-27T16:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.768892 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.768957 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.768969 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.768989 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.769001 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:40Z","lastTransitionTime":"2026-02-27T16:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.871937 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.872001 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.872018 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.872046 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.872064 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:40Z","lastTransitionTime":"2026-02-27T16:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.974305 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.974335 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.974342 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.974356 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:40 crc kubenswrapper[4814]: I0227 16:24:40.974365 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:40Z","lastTransitionTime":"2026-02-27T16:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.076951 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.077037 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.077060 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.077091 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.077110 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:41Z","lastTransitionTime":"2026-02-27T16:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.156382 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.156471 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.156513 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.156532 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.156549 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:41 crc kubenswrapper[4814]: E0227 16:24:41.156644 4814 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:24:41 crc kubenswrapper[4814]: E0227 16:24:41.156692 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:49.156678835 +0000 UTC m=+101.609303665 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:24:41 crc kubenswrapper[4814]: E0227 16:24:41.156787 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:24:49.156702686 +0000 UTC m=+101.609327546 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:24:41 crc kubenswrapper[4814]: E0227 16:24:41.157026 4814 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:24:41 crc kubenswrapper[4814]: E0227 16:24:41.157039 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:24:41 crc kubenswrapper[4814]: E0227 16:24:41.157076 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:24:41 crc kubenswrapper[4814]: E0227 16:24:41.157101 4814 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:41 crc kubenswrapper[4814]: E0227 16:24:41.157125 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:24:41 crc kubenswrapper[4814]: E0227 16:24:41.157139 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:24:41 crc kubenswrapper[4814]: E0227 16:24:41.157150 4814 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:41 crc kubenswrapper[4814]: E0227 16:24:41.157052 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:49.157045678 +0000 UTC m=+101.609670508 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:24:41 crc kubenswrapper[4814]: E0227 16:24:41.157189 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:49.157179773 +0000 UTC m=+101.609804603 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:41 crc kubenswrapper[4814]: E0227 16:24:41.157200 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 16:24:49.157194943 +0000 UTC m=+101.609819773 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.179860 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.179891 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.179902 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.179916 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.179925 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:41Z","lastTransitionTime":"2026-02-27T16:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.283826 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.283896 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.283929 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.283960 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.283980 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:41Z","lastTransitionTime":"2026-02-27T16:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.386939 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.387012 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.387054 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.387077 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.387092 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:41Z","lastTransitionTime":"2026-02-27T16:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.487514 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.487885 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:41 crc kubenswrapper[4814]: E0227 16:24:41.488136 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.488286 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:41 crc kubenswrapper[4814]: E0227 16:24:41.488375 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:24:41 crc kubenswrapper[4814]: E0227 16:24:41.487898 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.490503 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.490616 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.490697 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.490785 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.490874 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:41Z","lastTransitionTime":"2026-02-27T16:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.593147 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.593244 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.593473 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.593505 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.593526 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:41Z","lastTransitionTime":"2026-02-27T16:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.697329 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.697402 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.697420 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.697450 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.697474 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:41Z","lastTransitionTime":"2026-02-27T16:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.801478 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.801555 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.801574 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.801602 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.801624 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:41Z","lastTransitionTime":"2026-02-27T16:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.904232 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.904330 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.904351 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.904376 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:41 crc kubenswrapper[4814]: I0227 16:24:41.904393 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:41Z","lastTransitionTime":"2026-02-27T16:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.007942 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.008012 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.008030 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.008056 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.008073 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:42Z","lastTransitionTime":"2026-02-27T16:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.111381 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.111441 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.111461 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.111484 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.111503 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:42Z","lastTransitionTime":"2026-02-27T16:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.214136 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.214190 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.214206 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.214228 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.214245 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:42Z","lastTransitionTime":"2026-02-27T16:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.317541 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.317609 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.317625 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.317654 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.317671 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:42Z","lastTransitionTime":"2026-02-27T16:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.420155 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.420203 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.420222 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.420240 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.420282 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:42Z","lastTransitionTime":"2026-02-27T16:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.523582 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.523636 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.523653 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.523678 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.523694 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:42Z","lastTransitionTime":"2026-02-27T16:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.627027 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.627084 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.627097 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.627117 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.627132 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:42Z","lastTransitionTime":"2026-02-27T16:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.729481 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.729551 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.729569 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.729595 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.729614 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:42Z","lastTransitionTime":"2026-02-27T16:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.832179 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.832219 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.832238 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.832297 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.832314 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:42Z","lastTransitionTime":"2026-02-27T16:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.936014 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.936069 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.936090 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.936116 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:42 crc kubenswrapper[4814]: I0227 16:24:42.936135 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:42Z","lastTransitionTime":"2026-02-27T16:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.038776 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.038828 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.038845 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.038870 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.038888 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:43Z","lastTransitionTime":"2026-02-27T16:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.141338 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.141402 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.141420 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.141442 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.141459 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:43Z","lastTransitionTime":"2026-02-27T16:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.244503 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.244560 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.244576 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.245365 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.245386 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:43Z","lastTransitionTime":"2026-02-27T16:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.348551 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.348636 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.348655 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.348679 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.348696 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:43Z","lastTransitionTime":"2026-02-27T16:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.452072 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.452115 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.452124 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.452139 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.452149 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:43Z","lastTransitionTime":"2026-02-27T16:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.487154 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.487224 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:43 crc kubenswrapper[4814]: E0227 16:24:43.487293 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:24:43 crc kubenswrapper[4814]: E0227 16:24:43.487399 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.487457 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:43 crc kubenswrapper[4814]: E0227 16:24:43.487540 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.555159 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.555206 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.555218 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.555239 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.555270 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:43Z","lastTransitionTime":"2026-02-27T16:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.657577 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.657640 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.657657 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.657687 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.657708 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:43Z","lastTransitionTime":"2026-02-27T16:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.759852 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.759919 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.759936 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.759961 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.759982 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:43Z","lastTransitionTime":"2026-02-27T16:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.863377 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.863447 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.863464 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.863490 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.863508 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:43Z","lastTransitionTime":"2026-02-27T16:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.969372 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.969444 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.969463 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.969490 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:43 crc kubenswrapper[4814]: I0227 16:24:43.969510 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:43Z","lastTransitionTime":"2026-02-27T16:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.073666 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.074125 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.074276 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.074458 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.074615 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:44Z","lastTransitionTime":"2026-02-27T16:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.178186 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.178285 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.178309 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.178337 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.178361 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:44Z","lastTransitionTime":"2026-02-27T16:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.281172 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.281250 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.281310 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.281340 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.281358 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:44Z","lastTransitionTime":"2026-02-27T16:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.384713 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.384766 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.384779 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.384796 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.384810 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:44Z","lastTransitionTime":"2026-02-27T16:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.487470 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.487539 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.487558 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.487580 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.487602 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:44Z","lastTransitionTime":"2026-02-27T16:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.590962 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.591043 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.591069 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.591102 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.591125 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:44Z","lastTransitionTime":"2026-02-27T16:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.693893 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.693980 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.694010 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.694031 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.694044 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:44Z","lastTransitionTime":"2026-02-27T16:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.796593 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.796657 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.796677 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.796713 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.796732 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:44Z","lastTransitionTime":"2026-02-27T16:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.899003 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.899058 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.899077 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.899099 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:44 crc kubenswrapper[4814]: I0227 16:24:44.899118 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:44Z","lastTransitionTime":"2026-02-27T16:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.002491 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.002566 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.002583 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.002608 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.002629 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:45Z","lastTransitionTime":"2026-02-27T16:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.105508 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.105576 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.105593 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.105618 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.105636 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:45Z","lastTransitionTime":"2026-02-27T16:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.208826 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.208937 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.209018 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.209086 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.209143 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:45Z","lastTransitionTime":"2026-02-27T16:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.312269 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.312486 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.312572 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.312761 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.312841 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:45Z","lastTransitionTime":"2026-02-27T16:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.416183 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.416486 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.416643 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.416784 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.416929 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:45Z","lastTransitionTime":"2026-02-27T16:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.487081 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.487174 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:45 crc kubenswrapper[4814]: E0227 16:24:45.487378 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.487582 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:45 crc kubenswrapper[4814]: E0227 16:24:45.487587 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:24:45 crc kubenswrapper[4814]: E0227 16:24:45.487985 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.520455 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.520531 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.520549 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.520580 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.520604 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:45Z","lastTransitionTime":"2026-02-27T16:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.534545 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.534660 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.534689 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.534723 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.534752 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:45Z","lastTransitionTime":"2026-02-27T16:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:45 crc kubenswrapper[4814]: E0227 16:24:45.557804 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:45Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.563652 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.563705 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.563729 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.563761 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.563785 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:45Z","lastTransitionTime":"2026-02-27T16:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:45 crc kubenswrapper[4814]: E0227 16:24:45.581486 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:45Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.587208 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.587300 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.587315 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.587341 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.587361 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:45Z","lastTransitionTime":"2026-02-27T16:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:45 crc kubenswrapper[4814]: E0227 16:24:45.604911 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:45Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.610387 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.610647 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.610838 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.611082 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.611310 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:45Z","lastTransitionTime":"2026-02-27T16:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:45 crc kubenswrapper[4814]: E0227 16:24:45.630828 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:45Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.637131 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.637185 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.637205 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.637234 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.637279 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:45Z","lastTransitionTime":"2026-02-27T16:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:45 crc kubenswrapper[4814]: E0227 16:24:45.653303 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:45Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:45 crc kubenswrapper[4814]: E0227 16:24:45.653488 4814 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.655934 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.655984 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.656002 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.656024 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.656040 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:45Z","lastTransitionTime":"2026-02-27T16:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.759976 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.760043 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.760063 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.760089 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.760110 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:45Z","lastTransitionTime":"2026-02-27T16:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.863175 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.863230 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.863248 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.863299 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.863317 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:45Z","lastTransitionTime":"2026-02-27T16:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.966441 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.966518 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.966536 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.966558 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:45 crc kubenswrapper[4814]: I0227 16:24:45.966575 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:45Z","lastTransitionTime":"2026-02-27T16:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.070691 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.070958 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.071146 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.071367 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.071516 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:46Z","lastTransitionTime":"2026-02-27T16:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.175051 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.175116 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.175133 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.175159 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.175182 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:46Z","lastTransitionTime":"2026-02-27T16:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.279432 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.279517 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.279539 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.279567 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.279588 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:46Z","lastTransitionTime":"2026-02-27T16:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.383119 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.383214 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.383247 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.383325 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.383351 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:46Z","lastTransitionTime":"2026-02-27T16:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.485871 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.485997 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.486019 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.486051 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.486071 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:46Z","lastTransitionTime":"2026-02-27T16:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.588730 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.588984 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.589138 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.589382 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.589554 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:46Z","lastTransitionTime":"2026-02-27T16:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.695089 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.695440 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.695617 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.695760 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.695880 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:46Z","lastTransitionTime":"2026-02-27T16:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.800394 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.800455 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.800472 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.800500 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.800518 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:46Z","lastTransitionTime":"2026-02-27T16:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.903158 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.903218 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.903236 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.903289 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:46 crc kubenswrapper[4814]: I0227 16:24:46.903309 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:46Z","lastTransitionTime":"2026-02-27T16:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.006526 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.007327 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.007518 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.007709 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.007874 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:47Z","lastTransitionTime":"2026-02-27T16:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.110359 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.110420 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.110444 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.110474 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.110497 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:47Z","lastTransitionTime":"2026-02-27T16:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.213141 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.213191 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.213209 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.213232 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.213249 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:47Z","lastTransitionTime":"2026-02-27T16:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.317186 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.317722 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.317937 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.318132 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.318329 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:47Z","lastTransitionTime":"2026-02-27T16:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.421069 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.421156 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.421176 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.421207 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.421229 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:47Z","lastTransitionTime":"2026-02-27T16:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.487479 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.487589 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.487845 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:47 crc kubenswrapper[4814]: E0227 16:24:47.488145 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:24:47 crc kubenswrapper[4814]: E0227 16:24:47.488607 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:24:47 crc kubenswrapper[4814]: E0227 16:24:47.488769 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.524935 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.525016 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.525037 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.525071 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.525094 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:47Z","lastTransitionTime":"2026-02-27T16:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.628839 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.629295 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.629433 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.629623 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.629774 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:47Z","lastTransitionTime":"2026-02-27T16:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.733235 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.733795 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.733942 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.734100 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.734277 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:47Z","lastTransitionTime":"2026-02-27T16:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.837972 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.838052 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.838079 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.838112 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.838138 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:47Z","lastTransitionTime":"2026-02-27T16:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.941100 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.941180 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.941200 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.941232 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:47 crc kubenswrapper[4814]: I0227 16:24:47.941280 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:47Z","lastTransitionTime":"2026-02-27T16:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.045072 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.045138 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.045157 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.045184 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.045204 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:48Z","lastTransitionTime":"2026-02-27T16:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.148439 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.148505 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.148525 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.148553 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.148573 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:48Z","lastTransitionTime":"2026-02-27T16:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.251763 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.251842 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.251861 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.251890 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.251917 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:48Z","lastTransitionTime":"2026-02-27T16:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.355721 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.355786 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.355804 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.355829 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.355849 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:48Z","lastTransitionTime":"2026-02-27T16:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.459499 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.459562 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.459579 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.459606 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.459624 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:48Z","lastTransitionTime":"2026-02-27T16:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.506938 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.527031 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.545922 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.563139 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.563202 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.563220 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.563246 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.563303 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:48Z","lastTransitionTime":"2026-02-27T16:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.569570 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.589567 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.620055 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.645719 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.666103 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.666161 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.666173 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.666194 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.666211 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:48Z","lastTransitionTime":"2026-02-27T16:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.769095 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.769154 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.769165 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.769192 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.769205 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:48Z","lastTransitionTime":"2026-02-27T16:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.872242 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.872334 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.872357 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.872381 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.872399 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:48Z","lastTransitionTime":"2026-02-27T16:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.975714 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.975779 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.975792 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.975810 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:48 crc kubenswrapper[4814]: I0227 16:24:48.975824 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:48Z","lastTransitionTime":"2026-02-27T16:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.079122 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.079211 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.079233 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.079302 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.079325 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:49Z","lastTransitionTime":"2026-02-27T16:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.183468 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.183540 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.183561 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.183594 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.183616 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:49Z","lastTransitionTime":"2026-02-27T16:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.240008 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.240193 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:49 crc kubenswrapper[4814]: E0227 16:24:49.240227 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:25:05.240195134 +0000 UTC m=+117.692819974 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.240311 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.240359 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.240397 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:49 crc kubenswrapper[4814]: E0227 16:24:49.240465 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:24:49 crc kubenswrapper[4814]: E0227 16:24:49.240525 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:24:49 crc kubenswrapper[4814]: E0227 16:24:49.240543 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:24:49 crc kubenswrapper[4814]: E0227 16:24:49.240549 4814 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:49 crc kubenswrapper[4814]: E0227 16:24:49.240563 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:24:49 crc kubenswrapper[4814]: E0227 16:24:49.240564 4814 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:24:49 crc kubenswrapper[4814]: E0227 16:24:49.240609 4814 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:24:49 crc kubenswrapper[4814]: E0227 16:24:49.240645 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 16:25:05.240620459 +0000 UTC m=+117.693245319 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:49 crc kubenswrapper[4814]: E0227 16:24:49.240578 4814 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:49 crc kubenswrapper[4814]: E0227 16:24:49.240677 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:25:05.24066337 +0000 UTC m=+117.693288240 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:24:49 crc kubenswrapper[4814]: E0227 16:24:49.240699 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:25:05.240688171 +0000 UTC m=+117.693313041 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:24:49 crc kubenswrapper[4814]: E0227 16:24:49.240731 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 16:25:05.240719252 +0000 UTC m=+117.693344112 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.289896 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.289978 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.290008 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.290044 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.290079 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:49Z","lastTransitionTime":"2026-02-27T16:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.394049 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.394117 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.394137 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.394163 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.394186 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:49Z","lastTransitionTime":"2026-02-27T16:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.487475 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.487608 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:49 crc kubenswrapper[4814]: E0227 16:24:49.487659 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.488191 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:49 crc kubenswrapper[4814]: E0227 16:24:49.489950 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:24:49 crc kubenswrapper[4814]: E0227 16:24:49.490066 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.498160 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.498235 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.498285 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.498316 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.498339 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:49Z","lastTransitionTime":"2026-02-27T16:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.508606 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.601336 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.601389 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.601409 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.601432 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.601450 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:49Z","lastTransitionTime":"2026-02-27T16:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.705561 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.705624 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.705644 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.705670 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.705693 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:49Z","lastTransitionTime":"2026-02-27T16:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.809311 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.809360 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.809377 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.809400 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.809416 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:49Z","lastTransitionTime":"2026-02-27T16:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.912412 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.912482 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.912502 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.912530 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:49 crc kubenswrapper[4814]: I0227 16:24:49.912549 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:49Z","lastTransitionTime":"2026-02-27T16:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.016996 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.017562 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.017931 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.018147 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.018378 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:50Z","lastTransitionTime":"2026-02-27T16:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.122678 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.123765 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.123926 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.124101 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.124291 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:50Z","lastTransitionTime":"2026-02-27T16:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.227644 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.227984 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.228441 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.228737 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.229031 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:50Z","lastTransitionTime":"2026-02-27T16:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.332556 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.332590 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.332599 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.332611 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.332622 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:50Z","lastTransitionTime":"2026-02-27T16:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.435875 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.435949 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.435967 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.435992 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.436009 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:50Z","lastTransitionTime":"2026-02-27T16:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.488361 4814 scope.go:117] "RemoveContainer" containerID="bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1" Feb 27 16:24:50 crc kubenswrapper[4814]: E0227 16:24:50.488585 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.539404 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.539482 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.539506 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.539531 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.539549 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:50Z","lastTransitionTime":"2026-02-27T16:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.644450 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.644547 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.644575 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.644614 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.644638 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:50Z","lastTransitionTime":"2026-02-27T16:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.748356 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.748535 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.748591 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.748627 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.748648 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:50Z","lastTransitionTime":"2026-02-27T16:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.852186 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.852292 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.852314 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.852343 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.852365 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:50Z","lastTransitionTime":"2026-02-27T16:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.955171 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.955486 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.955656 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.955981 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:50 crc kubenswrapper[4814]: I0227 16:24:50.956121 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:50Z","lastTransitionTime":"2026-02-27T16:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.058702 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.058764 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.058782 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.058807 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.058826 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:51Z","lastTransitionTime":"2026-02-27T16:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.161640 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.161703 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.161723 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.161828 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.161849 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:51Z","lastTransitionTime":"2026-02-27T16:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.265246 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.265350 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.265368 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.265392 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.265412 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:51Z","lastTransitionTime":"2026-02-27T16:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.369610 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.369679 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.369697 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.369724 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.369743 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:51Z","lastTransitionTime":"2026-02-27T16:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.472954 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.473026 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.473046 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.473075 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.473096 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:51Z","lastTransitionTime":"2026-02-27T16:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.486770 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.486775 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.486854 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:51 crc kubenswrapper[4814]: E0227 16:24:51.486946 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:24:51 crc kubenswrapper[4814]: E0227 16:24:51.487193 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:24:51 crc kubenswrapper[4814]: E0227 16:24:51.487517 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.576598 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.576686 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.576711 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.576751 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.576777 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:51Z","lastTransitionTime":"2026-02-27T16:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.679786 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.679861 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.679882 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.679915 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.679953 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:51Z","lastTransitionTime":"2026-02-27T16:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.782335 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.782379 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.782390 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.782408 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.782420 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:51Z","lastTransitionTime":"2026-02-27T16:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.885497 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.885565 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.885581 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.885606 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.885623 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:51Z","lastTransitionTime":"2026-02-27T16:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.988500 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.988915 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.989013 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.989106 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:51 crc kubenswrapper[4814]: I0227 16:24:51.989193 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:51Z","lastTransitionTime":"2026-02-27T16:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.092812 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.092874 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.092893 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.092918 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.092937 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:52Z","lastTransitionTime":"2026-02-27T16:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.165429 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-hmxvd"] Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.166104 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hmxvd" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.170144 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9e61af0a-b3e1-4131-a0d4-2c550b00cb16-hosts-file\") pod \"node-resolver-hmxvd\" (UID: \"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\") " pod="openshift-dns/node-resolver-hmxvd" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.170234 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9zkk\" (UniqueName: \"kubernetes.io/projected/9e61af0a-b3e1-4131-a0d4-2c550b00cb16-kube-api-access-q9zkk\") pod \"node-resolver-hmxvd\" (UID: \"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\") " pod="openshift-dns/node-resolver-hmxvd" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.170849 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.171056 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.172655 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.197017 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.197084 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.197108 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.197141 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.197168 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:52Z","lastTransitionTime":"2026-02-27T16:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.198235 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.221957 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.237074 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.271233 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9e61af0a-b3e1-4131-a0d4-2c550b00cb16-hosts-file\") pod \"node-resolver-hmxvd\" (UID: \"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\") " pod="openshift-dns/node-resolver-hmxvd" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.271365 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9zkk\" (UniqueName: \"kubernetes.io/projected/9e61af0a-b3e1-4131-a0d4-2c550b00cb16-kube-api-access-q9zkk\") pod \"node-resolver-hmxvd\" (UID: \"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\") " pod="openshift-dns/node-resolver-hmxvd" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.271565 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9e61af0a-b3e1-4131-a0d4-2c550b00cb16-hosts-file\") pod \"node-resolver-hmxvd\" (UID: \"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\") " pod="openshift-dns/node-resolver-hmxvd" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.278081 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.301462 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.301530 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.301549 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.301576 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.301594 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:52Z","lastTransitionTime":"2026-02-27T16:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.309310 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9zkk\" (UniqueName: \"kubernetes.io/projected/9e61af0a-b3e1-4131-a0d4-2c550b00cb16-kube-api-access-q9zkk\") pod \"node-resolver-hmxvd\" (UID: \"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\") " pod="openshift-dns/node-resolver-hmxvd" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.311734 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.336098 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.358244 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.382064 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.402755 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.405600 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.405672 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.405691 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.405718 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.405736 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:52Z","lastTransitionTime":"2026-02-27T16:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.492795 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hmxvd" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.509938 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.510174 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.510349 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.510512 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.510644 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:52Z","lastTransitionTime":"2026-02-27T16:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:52 crc kubenswrapper[4814]: W0227 16:24:52.518023 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e61af0a_b3e1_4131_a0d4_2c550b00cb16.slice/crio-3837848c1f2ebc20f2cfcf14c128c11023849630f2c6cd7699437472035500bc WatchSource:0}: Error finding container 3837848c1f2ebc20f2cfcf14c128c11023849630f2c6cd7699437472035500bc: Status 404 returned error can't find the container with id 3837848c1f2ebc20f2cfcf14c128c11023849630f2c6cd7699437472035500bc Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.566245 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-nlbk2"] Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.567406 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.569857 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-sh8gf"] Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.570500 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.571043 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.571726 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-4m267"] Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.572089 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.572113 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.572533 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.573088 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.572184 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574024 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-multus-socket-dir-parent\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574116 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6thr9\" (UniqueName: \"kubernetes.io/projected/be713cf4-11a1-42e7-93c6-ab3f748570a4-kube-api-access-6thr9\") pod \"machine-config-daemon-nlbk2\" (UID: \"be713cf4-11a1-42e7-93c6-ab3f748570a4\") " pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574173 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-var-lib-cni-multus\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574221 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2bf0649e-4683-4705-9006-df485b9b5250-os-release\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574306 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2bf0649e-4683-4705-9006-df485b9b5250-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574356 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-os-release\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574401 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/be713cf4-11a1-42e7-93c6-ab3f748570a4-mcd-auth-proxy-config\") pod \"machine-config-daemon-nlbk2\" (UID: \"be713cf4-11a1-42e7-93c6-ab3f748570a4\") " pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574449 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/56739684-228b-4bf3-bb4e-11f7aaca1890-multus-daemon-config\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574495 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/be713cf4-11a1-42e7-93c6-ab3f748570a4-rootfs\") pod \"machine-config-daemon-nlbk2\" (UID: \"be713cf4-11a1-42e7-93c6-ab3f748570a4\") " pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574540 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ms7n\" (UniqueName: \"kubernetes.io/projected/2bf0649e-4683-4705-9006-df485b9b5250-kube-api-access-9ms7n\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574586 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-multus-cni-dir\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574630 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-var-lib-kubelet\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574750 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-system-cni-dir\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574786 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574803 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/56739684-228b-4bf3-bb4e-11f7aaca1890-cni-binary-copy\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574866 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-etc-kubernetes\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574914 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-run-k8s-cni-cncf-io\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.574959 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-cnibin\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.575005 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdjfb\" (UniqueName: \"kubernetes.io/projected/56739684-228b-4bf3-bb4e-11f7aaca1890-kube-api-access-kdjfb\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.575050 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/be713cf4-11a1-42e7-93c6-ab3f748570a4-proxy-tls\") pod \"machine-config-daemon-nlbk2\" (UID: \"be713cf4-11a1-42e7-93c6-ab3f748570a4\") " pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.575097 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2bf0649e-4683-4705-9006-df485b9b5250-cni-binary-copy\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.575148 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-run-netns\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.575191 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2bf0649e-4683-4705-9006-df485b9b5250-cnibin\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.575239 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2bf0649e-4683-4705-9006-df485b9b5250-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.575330 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-run-multus-certs\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.575402 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-var-lib-cni-bin\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.575457 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-multus-conf-dir\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.575508 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2bf0649e-4683-4705-9006-df485b9b5250-system-cni-dir\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.575711 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-hostroot\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.575803 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.576184 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.576644 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.576762 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.576650 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.577825 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.590834 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.611609 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.615366 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.615419 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.615438 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.615471 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.615491 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:52Z","lastTransitionTime":"2026-02-27T16:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.631785 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.648737 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.665979 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.676783 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-etc-kubernetes\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.676854 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-run-k8s-cni-cncf-io\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.676887 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2bf0649e-4683-4705-9006-df485b9b5250-cni-binary-copy\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.676927 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-cnibin\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.676959 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdjfb\" (UniqueName: \"kubernetes.io/projected/56739684-228b-4bf3-bb4e-11f7aaca1890-kube-api-access-kdjfb\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.676989 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/be713cf4-11a1-42e7-93c6-ab3f748570a4-proxy-tls\") pod \"machine-config-daemon-nlbk2\" (UID: \"be713cf4-11a1-42e7-93c6-ab3f748570a4\") " pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677020 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-run-netns\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677052 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2bf0649e-4683-4705-9006-df485b9b5250-cnibin\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677085 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2bf0649e-4683-4705-9006-df485b9b5250-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677172 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-run-multus-certs\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677208 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-var-lib-cni-bin\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677238 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-multus-conf-dir\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677298 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2bf0649e-4683-4705-9006-df485b9b5250-system-cni-dir\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677332 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-hostroot\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677361 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-multus-socket-dir-parent\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677395 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6thr9\" (UniqueName: \"kubernetes.io/projected/be713cf4-11a1-42e7-93c6-ab3f748570a4-kube-api-access-6thr9\") pod \"machine-config-daemon-nlbk2\" (UID: \"be713cf4-11a1-42e7-93c6-ab3f748570a4\") " pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677428 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-var-lib-cni-multus\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677459 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2bf0649e-4683-4705-9006-df485b9b5250-os-release\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677488 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2bf0649e-4683-4705-9006-df485b9b5250-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677521 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-os-release\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677548 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/be713cf4-11a1-42e7-93c6-ab3f748570a4-mcd-auth-proxy-config\") pod \"machine-config-daemon-nlbk2\" (UID: \"be713cf4-11a1-42e7-93c6-ab3f748570a4\") " pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677579 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-multus-cni-dir\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677609 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-var-lib-kubelet\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677636 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-cnibin\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677640 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/56739684-228b-4bf3-bb4e-11f7aaca1890-multus-daemon-config\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677716 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/be713cf4-11a1-42e7-93c6-ab3f748570a4-rootfs\") pod \"machine-config-daemon-nlbk2\" (UID: \"be713cf4-11a1-42e7-93c6-ab3f748570a4\") " pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677746 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ms7n\" (UniqueName: \"kubernetes.io/projected/2bf0649e-4683-4705-9006-df485b9b5250-kube-api-access-9ms7n\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677807 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-system-cni-dir\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.677828 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/56739684-228b-4bf3-bb4e-11f7aaca1890-cni-binary-copy\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678223 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-run-netns\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678359 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-os-release\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678450 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-var-lib-kubelet\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678513 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-var-lib-cni-multus\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678527 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2bf0649e-4683-4705-9006-df485b9b5250-cnibin\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678460 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2bf0649e-4683-4705-9006-df485b9b5250-os-release\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678594 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-multus-cni-dir\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678632 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-multus-socket-dir-parent\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678704 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-hostroot\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678713 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-var-lib-cni-bin\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678742 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-run-multus-certs\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678776 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-etc-kubernetes\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678838 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-host-run-k8s-cni-cncf-io\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678848 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-multus-conf-dir\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678861 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/be713cf4-11a1-42e7-93c6-ab3f748570a4-rootfs\") pod \"machine-config-daemon-nlbk2\" (UID: \"be713cf4-11a1-42e7-93c6-ab3f748570a4\") " pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678848 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/56739684-228b-4bf3-bb4e-11f7aaca1890-system-cni-dir\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678910 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/56739684-228b-4bf3-bb4e-11f7aaca1890-multus-daemon-config\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.678414 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2bf0649e-4683-4705-9006-df485b9b5250-system-cni-dir\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.679985 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/be713cf4-11a1-42e7-93c6-ab3f748570a4-mcd-auth-proxy-config\") pod \"machine-config-daemon-nlbk2\" (UID: \"be713cf4-11a1-42e7-93c6-ab3f748570a4\") " pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.682796 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2bf0649e-4683-4705-9006-df485b9b5250-cni-binary-copy\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.683384 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2bf0649e-4683-4705-9006-df485b9b5250-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.683764 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/56739684-228b-4bf3-bb4e-11f7aaca1890-cni-binary-copy\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.689928 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/be713cf4-11a1-42e7-93c6-ab3f748570a4-proxy-tls\") pod \"machine-config-daemon-nlbk2\" (UID: \"be713cf4-11a1-42e7-93c6-ab3f748570a4\") " pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.698530 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2bf0649e-4683-4705-9006-df485b9b5250-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.703041 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdjfb\" (UniqueName: \"kubernetes.io/projected/56739684-228b-4bf3-bb4e-11f7aaca1890-kube-api-access-kdjfb\") pod \"multus-4m267\" (UID: \"56739684-228b-4bf3-bb4e-11f7aaca1890\") " pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.707405 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6thr9\" (UniqueName: \"kubernetes.io/projected/be713cf4-11a1-42e7-93c6-ab3f748570a4-kube-api-access-6thr9\") pod \"machine-config-daemon-nlbk2\" (UID: \"be713cf4-11a1-42e7-93c6-ab3f748570a4\") " pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.707328 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.712861 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ms7n\" (UniqueName: \"kubernetes.io/projected/2bf0649e-4683-4705-9006-df485b9b5250-kube-api-access-9ms7n\") pod \"multus-additional-cni-plugins-sh8gf\" (UID: \"2bf0649e-4683-4705-9006-df485b9b5250\") " pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.718200 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.718296 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.718321 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.718351 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.718371 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:52Z","lastTransitionTime":"2026-02-27T16:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.726928 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.747907 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.768453 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.783917 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.800475 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.818442 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.821578 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.821614 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.821629 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.821652 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.821668 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:52Z","lastTransitionTime":"2026-02-27T16:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.841652 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.859985 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.879710 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.895613 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.901092 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.914479 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4m267" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.915329 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.919153 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hmxvd" event={"ID":"9e61af0a-b3e1-4131-a0d4-2c550b00cb16","Type":"ContainerStarted","Data":"3837848c1f2ebc20f2cfcf14c128c11023849630f2c6cd7699437472035500bc"} Feb 27 16:24:52 crc kubenswrapper[4814]: W0227 16:24:52.920969 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe713cf4_11a1_42e7_93c6_ab3f748570a4.slice/crio-ba55ab5d5f817c3209b95e25be61c2bb81474273546a62829528683bfcd14b4c WatchSource:0}: Error finding container ba55ab5d5f817c3209b95e25be61c2bb81474273546a62829528683bfcd14b4c: Status 404 returned error can't find the container with id ba55ab5d5f817c3209b95e25be61c2bb81474273546a62829528683bfcd14b4c Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.923205 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.923228 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.923238 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.923278 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.923294 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:52Z","lastTransitionTime":"2026-02-27T16:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.925236 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" Feb 27 16:24:52 crc kubenswrapper[4814]: W0227 16:24:52.933733 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56739684_228b_4bf3_bb4e_11f7aaca1890.slice/crio-42b2566da26be212d7d42207477becdb1b8bcb053987f3b7e1d84bb0f82c7a34 WatchSource:0}: Error finding container 42b2566da26be212d7d42207477becdb1b8bcb053987f3b7e1d84bb0f82c7a34: Status 404 returned error can't find the container with id 42b2566da26be212d7d42207477becdb1b8bcb053987f3b7e1d84bb0f82c7a34 Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.936549 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.949921 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.969536 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.971942 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7zrr2"] Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.972832 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.975455 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.975656 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.976122 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.976244 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.976452 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.976357 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.976581 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 27 16:24:52 crc kubenswrapper[4814]: W0227 16:24:52.978269 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bf0649e_4683_4705_9006_df485b9b5250.slice/crio-c3894ac108b0dd0420045c32ed12986932d6f9c59aabe196eaf0f11a0c8ec882 WatchSource:0}: Error finding container c3894ac108b0dd0420045c32ed12986932d6f9c59aabe196eaf0f11a0c8ec882: Status 404 returned error can't find the container with id c3894ac108b0dd0420045c32ed12986932d6f9c59aabe196eaf0f11a0c8ec882 Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.980770 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-cni-netd\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.980838 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-slash\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.980874 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-etc-openvswitch\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.980906 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-cni-bin\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.980939 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-kubelet\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.980975 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-systemd-units\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.981006 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-ovn\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.981042 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-log-socket\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.981077 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-run-ovn-kubernetes\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.981162 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-ovnkube-config\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.981196 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jntcv\" (UniqueName: \"kubernetes.io/projected/4619e86c-7eae-4849-89b0-abb44778d871-kube-api-access-jntcv\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.981301 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-openvswitch\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.981348 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-run-netns\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.981383 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-var-lib-openvswitch\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.981419 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-env-overrides\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.981450 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-ovnkube-script-lib\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.981485 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.981548 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4619e86c-7eae-4849-89b0-abb44778d871-ovn-node-metrics-cert\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.981600 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-systemd\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.981634 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-node-log\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:52 crc kubenswrapper[4814]: I0227 16:24:52.997417 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:52Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.016437 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.025977 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.026014 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.026027 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.026046 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.026058 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:53Z","lastTransitionTime":"2026-02-27T16:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.047527 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.071386 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.083873 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-openvswitch\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.083953 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-run-netns\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.083991 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-var-lib-openvswitch\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084026 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-ovnkube-script-lib\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084060 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-env-overrides\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084094 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-systemd\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084105 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-openvswitch\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084172 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-node-log\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084217 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084241 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-var-lib-openvswitch\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084275 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4619e86c-7eae-4849-89b0-abb44778d871-ovn-node-metrics-cert\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084407 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-slash\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084484 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-etc-openvswitch\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084581 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-slash\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084522 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-cni-bin\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084853 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-etc-openvswitch\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084945 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-env-overrides\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084638 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-node-log\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084175 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-run-netns\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084701 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084790 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-cni-bin\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.084676 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-systemd\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.085049 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-cni-netd\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.085127 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-kubelet\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.085142 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-cni-netd\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.085196 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-systemd-units\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.085243 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-systemd-units\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.085302 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-ovn\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.085377 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-ovn\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.085396 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-ovnkube-script-lib\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.085529 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-log-socket\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.085641 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-run-ovn-kubernetes\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.085709 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-log-socket\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.085717 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-ovnkube-config\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.085813 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jntcv\" (UniqueName: \"kubernetes.io/projected/4619e86c-7eae-4849-89b0-abb44778d871-kube-api-access-jntcv\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.085811 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-run-ovn-kubernetes\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.086110 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-kubelet\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.086401 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-ovnkube-config\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.086413 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.090606 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4619e86c-7eae-4849-89b0-abb44778d871-ovn-node-metrics-cert\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.121050 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jntcv\" (UniqueName: \"kubernetes.io/projected/4619e86c-7eae-4849-89b0-abb44778d871-kube-api-access-jntcv\") pod \"ovnkube-node-7zrr2\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.127303 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.129600 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.129652 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.129663 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.129685 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.129730 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:53Z","lastTransitionTime":"2026-02-27T16:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.145116 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.165381 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.181937 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.202722 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.222293 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.232114 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.232161 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.232177 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.232200 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.232214 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:53Z","lastTransitionTime":"2026-02-27T16:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.240890 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.256005 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.271760 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.291239 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.296771 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.338908 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.338955 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.338970 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.338987 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.339000 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:53Z","lastTransitionTime":"2026-02-27T16:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.441451 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.441487 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.441499 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.441529 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.441543 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:53Z","lastTransitionTime":"2026-02-27T16:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.487423 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.487488 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.487543 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:53 crc kubenswrapper[4814]: E0227 16:24:53.487789 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:24:53 crc kubenswrapper[4814]: E0227 16:24:53.487644 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:24:53 crc kubenswrapper[4814]: E0227 16:24:53.487565 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.543862 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.543903 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.543915 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.543934 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.543947 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:53Z","lastTransitionTime":"2026-02-27T16:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.646725 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.646799 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.646817 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.646841 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.646860 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:53Z","lastTransitionTime":"2026-02-27T16:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.749631 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.749693 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.749710 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.749735 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.749753 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:53Z","lastTransitionTime":"2026-02-27T16:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.853066 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.853118 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.853138 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.853161 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.853179 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:53Z","lastTransitionTime":"2026-02-27T16:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.925887 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4m267" event={"ID":"56739684-228b-4bf3-bb4e-11f7aaca1890","Type":"ContainerStarted","Data":"46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.926143 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4m267" event={"ID":"56739684-228b-4bf3-bb4e-11f7aaca1890","Type":"ContainerStarted","Data":"42b2566da26be212d7d42207477becdb1b8bcb053987f3b7e1d84bb0f82c7a34"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.928181 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.928233 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.928281 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"ba55ab5d5f817c3209b95e25be61c2bb81474273546a62829528683bfcd14b4c"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.929803 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hmxvd" event={"ID":"9e61af0a-b3e1-4131-a0d4-2c550b00cb16","Type":"ContainerStarted","Data":"ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.931844 4814 generic.go:334] "Generic (PLEG): container finished" podID="4619e86c-7eae-4849-89b0-abb44778d871" containerID="5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42" exitCode=0 Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.931960 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerDied","Data":"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.932375 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerStarted","Data":"bd2b250dddc2ff4cf4bea73006bea42adf0ef0909bde6e8efef659d20f20f032"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.934134 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" event={"ID":"2bf0649e-4683-4705-9006-df485b9b5250","Type":"ContainerStarted","Data":"de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.934172 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" event={"ID":"2bf0649e-4683-4705-9006-df485b9b5250","Type":"ContainerStarted","Data":"c3894ac108b0dd0420045c32ed12986932d6f9c59aabe196eaf0f11a0c8ec882"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.952509 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.957915 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.958005 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.958044 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.958076 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.958097 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:53Z","lastTransitionTime":"2026-02-27T16:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:53 crc kubenswrapper[4814]: I0227 16:24:53.980511 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.000568 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:53Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.023941 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.042897 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.061615 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.061735 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.061748 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.061872 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.061885 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:54Z","lastTransitionTime":"2026-02-27T16:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.062220 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.096867 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.119735 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.139348 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.164867 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.165051 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.165113 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.165174 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.165229 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:54Z","lastTransitionTime":"2026-02-27T16:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.168386 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.191477 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.212672 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.229932 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.250352 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.269704 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.269756 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.269769 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.269785 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.269798 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:54Z","lastTransitionTime":"2026-02-27T16:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.281975 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.301880 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.327674 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.346769 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.362821 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.373104 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.373161 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.373180 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.373205 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.373229 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:54Z","lastTransitionTime":"2026-02-27T16:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.388549 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.409416 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.427348 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.451888 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.465542 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.475103 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.475165 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.475180 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.475198 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.475540 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:54Z","lastTransitionTime":"2026-02-27T16:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.491976 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.514226 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.578347 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.578399 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.578414 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.578435 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.578447 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:54Z","lastTransitionTime":"2026-02-27T16:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.681033 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.681072 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.681085 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.681103 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.681116 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:54Z","lastTransitionTime":"2026-02-27T16:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.784125 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.784207 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.784236 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.784734 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.784802 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:54Z","lastTransitionTime":"2026-02-27T16:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.887584 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.887616 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.887647 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.887659 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.887688 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:54Z","lastTransitionTime":"2026-02-27T16:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.943542 4814 generic.go:334] "Generic (PLEG): container finished" podID="2bf0649e-4683-4705-9006-df485b9b5250" containerID="de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304" exitCode=0 Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.943632 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" event={"ID":"2bf0649e-4683-4705-9006-df485b9b5250","Type":"ContainerDied","Data":"de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.952572 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerStarted","Data":"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.952616 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerStarted","Data":"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.952636 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerStarted","Data":"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.952652 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerStarted","Data":"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.952668 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerStarted","Data":"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.952684 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerStarted","Data":"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.963666 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.982036 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.990651 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.990685 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.990697 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.990715 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.990729 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:54Z","lastTransitionTime":"2026-02-27T16:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:54 crc kubenswrapper[4814]: I0227 16:24:54.996920 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.009867 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:55Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.021627 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:55Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.035636 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:55Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.053068 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:55Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.071486 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:55Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.100660 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.100697 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.100709 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.100725 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.100738 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:55Z","lastTransitionTime":"2026-02-27T16:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.102628 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:55Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.125574 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:55Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.152740 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:55Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.170198 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:55Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.186159 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:55Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.203738 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.203775 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.203787 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.203808 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.203821 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:55Z","lastTransitionTime":"2026-02-27T16:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.306416 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.306452 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.306465 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.306481 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.306494 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:55Z","lastTransitionTime":"2026-02-27T16:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.410273 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.410312 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.410320 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.410337 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.410347 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:55Z","lastTransitionTime":"2026-02-27T16:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.486554 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.486557 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.486684 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:55 crc kubenswrapper[4814]: E0227 16:24:55.486874 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:24:55 crc kubenswrapper[4814]: E0227 16:24:55.487171 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:24:55 crc kubenswrapper[4814]: E0227 16:24:55.487305 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.512369 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.512410 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.512420 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.512434 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.512445 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:55Z","lastTransitionTime":"2026-02-27T16:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.617832 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.617872 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.617883 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.617899 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.617910 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:55Z","lastTransitionTime":"2026-02-27T16:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.720081 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.720119 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.720131 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.720147 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.720160 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:55Z","lastTransitionTime":"2026-02-27T16:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.782083 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.782149 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.782172 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.782202 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.782227 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:55Z","lastTransitionTime":"2026-02-27T16:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:55 crc kubenswrapper[4814]: E0227 16:24:55.805492 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:55Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.812163 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.812224 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.812243 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.812301 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.812319 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:55Z","lastTransitionTime":"2026-02-27T16:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:55 crc kubenswrapper[4814]: E0227 16:24:55.834686 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:55Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.865212 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.865270 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.865279 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.865293 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.865303 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:55Z","lastTransitionTime":"2026-02-27T16:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:55 crc kubenswrapper[4814]: E0227 16:24:55.885446 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:55Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.889710 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.889745 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.889753 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.889770 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.889798 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:55Z","lastTransitionTime":"2026-02-27T16:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:55 crc kubenswrapper[4814]: E0227 16:24:55.906809 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:55Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.913847 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.914120 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.914160 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.914193 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.914217 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:55Z","lastTransitionTime":"2026-02-27T16:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:55 crc kubenswrapper[4814]: E0227 16:24:55.929392 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:55Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:55 crc kubenswrapper[4814]: E0227 16:24:55.929899 4814 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.932056 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.932103 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.932117 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.932135 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.932147 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:55Z","lastTransitionTime":"2026-02-27T16:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.957821 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" event={"ID":"2bf0649e-4683-4705-9006-df485b9b5250","Type":"ContainerStarted","Data":"8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b"} Feb 27 16:24:55 crc kubenswrapper[4814]: I0227 16:24:55.984209 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:55Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.004407 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:56Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.018821 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:56Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.034666 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.034704 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.034714 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.034730 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.034742 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:56Z","lastTransitionTime":"2026-02-27T16:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.040523 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:56Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.057552 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:56Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.078450 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:56Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.095549 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:56Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.118953 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:56Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.138840 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.138871 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.138882 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.138899 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.138911 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:56Z","lastTransitionTime":"2026-02-27T16:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.138991 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:56Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.156074 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:56Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.172100 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:56Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.188290 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:56Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.204733 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:56Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.242001 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.242080 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.242105 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.242137 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.242164 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:56Z","lastTransitionTime":"2026-02-27T16:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.344951 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.345022 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.345039 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.345057 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.345094 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:56Z","lastTransitionTime":"2026-02-27T16:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.448371 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.448470 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.448484 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.448502 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.448516 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:56Z","lastTransitionTime":"2026-02-27T16:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.551853 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.551917 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.551935 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.552137 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.552156 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:56Z","lastTransitionTime":"2026-02-27T16:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.655248 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.655338 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.655356 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.655378 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.655395 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:56Z","lastTransitionTime":"2026-02-27T16:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.758481 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.758545 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.758564 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.758591 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.758608 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:56Z","lastTransitionTime":"2026-02-27T16:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.861798 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.861857 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.861873 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.861895 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.861916 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:56Z","lastTransitionTime":"2026-02-27T16:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.965637 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.965698 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.965721 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.965719 4814 generic.go:334] "Generic (PLEG): container finished" podID="2bf0649e-4683-4705-9006-df485b9b5250" containerID="8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b" exitCode=0 Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.965750 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.965768 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" event={"ID":"2bf0649e-4683-4705-9006-df485b9b5250","Type":"ContainerDied","Data":"8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b"} Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.965772 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:56Z","lastTransitionTime":"2026-02-27T16:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:56 crc kubenswrapper[4814]: I0227 16:24:56.988664 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:56Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.010951 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:57Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.033375 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:57Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.051451 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:57Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.071323 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.071366 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.071382 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.071410 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.071432 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:57Z","lastTransitionTime":"2026-02-27T16:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.084665 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:57Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.108229 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:57Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.129572 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:57Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.151156 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:57Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.164624 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:57Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.179275 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.179333 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.179349 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.179375 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.179391 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:57Z","lastTransitionTime":"2026-02-27T16:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.189152 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:57Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.206897 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:57Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.229793 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:57Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.262298 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:57Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.281967 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.282024 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.282041 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.282064 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.282079 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:57Z","lastTransitionTime":"2026-02-27T16:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.385868 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.385904 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.385914 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.385929 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.385939 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:57Z","lastTransitionTime":"2026-02-27T16:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.487203 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.487231 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.487319 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:57 crc kubenswrapper[4814]: E0227 16:24:57.488088 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:24:57 crc kubenswrapper[4814]: E0227 16:24:57.488124 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:24:57 crc kubenswrapper[4814]: E0227 16:24:57.487913 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.490308 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.490405 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.490472 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.490499 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.490568 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:57Z","lastTransitionTime":"2026-02-27T16:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.594569 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.594631 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.594648 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.594673 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.594693 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:57Z","lastTransitionTime":"2026-02-27T16:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.697406 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.697491 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.697511 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.697566 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.697583 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:57Z","lastTransitionTime":"2026-02-27T16:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.801146 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.801205 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.801222 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.801245 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.801290 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:57Z","lastTransitionTime":"2026-02-27T16:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.903867 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.903926 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.903943 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.903963 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.903977 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:57Z","lastTransitionTime":"2026-02-27T16:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.977093 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerStarted","Data":"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1"} Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.980581 4814 generic.go:334] "Generic (PLEG): container finished" podID="2bf0649e-4683-4705-9006-df485b9b5250" containerID="0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde" exitCode=0 Feb 27 16:24:57 crc kubenswrapper[4814]: I0227 16:24:57.980642 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" event={"ID":"2bf0649e-4683-4705-9006-df485b9b5250","Type":"ContainerDied","Data":"0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde"} Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.007111 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.007172 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.007190 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.007215 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.007236 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:58Z","lastTransitionTime":"2026-02-27T16:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.037625 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.063067 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.078160 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.095944 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.109221 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.109292 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.109306 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.109323 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.109335 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:58Z","lastTransitionTime":"2026-02-27T16:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.113886 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.128833 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.143210 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.158027 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.172023 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.200046 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.212659 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.212713 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.212738 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.212767 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.212794 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:58Z","lastTransitionTime":"2026-02-27T16:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.221365 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.241643 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.274802 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.315125 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.315177 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.315195 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.315221 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.315240 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:58Z","lastTransitionTime":"2026-02-27T16:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.418375 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.418433 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.418452 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.418474 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.418500 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:58Z","lastTransitionTime":"2026-02-27T16:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.521206 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.521322 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.521344 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.521382 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.521412 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:58Z","lastTransitionTime":"2026-02-27T16:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.521975 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.546851 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.568406 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.600591 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.620163 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.624447 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.624495 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.624513 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.624536 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.624555 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:58Z","lastTransitionTime":"2026-02-27T16:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.645386 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.662859 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.684461 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.706941 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.728718 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.728812 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.728832 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.728867 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.728887 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:58Z","lastTransitionTime":"2026-02-27T16:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.734582 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.758134 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.777842 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.801441 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:58Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.831781 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.831877 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.831916 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.831971 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.831994 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:58Z","lastTransitionTime":"2026-02-27T16:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.935719 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.935797 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.935818 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.935845 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.935864 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:58Z","lastTransitionTime":"2026-02-27T16:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.990061 4814 generic.go:334] "Generic (PLEG): container finished" podID="2bf0649e-4683-4705-9006-df485b9b5250" containerID="5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c" exitCode=0 Feb 27 16:24:58 crc kubenswrapper[4814]: I0227 16:24:58.990126 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" event={"ID":"2bf0649e-4683-4705-9006-df485b9b5250","Type":"ContainerDied","Data":"5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c"} Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.015734 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.037426 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.039643 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.039707 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.039726 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.039756 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.039775 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:59Z","lastTransitionTime":"2026-02-27T16:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.055446 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.086234 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-5knkr"] Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.086799 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5knkr" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.086943 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.093754 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.093844 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.094051 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.093791 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.119621 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.141013 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.143010 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.143070 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.143088 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.143115 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.143134 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:59Z","lastTransitionTime":"2026-02-27T16:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.157600 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1c3aee15-1546-403a-866a-80455c2abb05-host\") pod \"node-ca-5knkr\" (UID: \"1c3aee15-1546-403a-866a-80455c2abb05\") " pod="openshift-image-registry/node-ca-5knkr" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.157683 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwhk2\" (UniqueName: \"kubernetes.io/projected/1c3aee15-1546-403a-866a-80455c2abb05-kube-api-access-xwhk2\") pod \"node-ca-5knkr\" (UID: \"1c3aee15-1546-403a-866a-80455c2abb05\") " pod="openshift-image-registry/node-ca-5knkr" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.157729 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1c3aee15-1546-403a-866a-80455c2abb05-serviceca\") pod \"node-ca-5knkr\" (UID: \"1c3aee15-1546-403a-866a-80455c2abb05\") " pod="openshift-image-registry/node-ca-5knkr" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.163050 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.182075 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.199104 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.238454 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.245887 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.245921 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.245932 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.245950 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.245962 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:59Z","lastTransitionTime":"2026-02-27T16:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.259045 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1c3aee15-1546-403a-866a-80455c2abb05-serviceca\") pod \"node-ca-5knkr\" (UID: \"1c3aee15-1546-403a-866a-80455c2abb05\") " pod="openshift-image-registry/node-ca-5knkr" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.259328 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1c3aee15-1546-403a-866a-80455c2abb05-host\") pod \"node-ca-5knkr\" (UID: \"1c3aee15-1546-403a-866a-80455c2abb05\") " pod="openshift-image-registry/node-ca-5knkr" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.259407 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwhk2\" (UniqueName: \"kubernetes.io/projected/1c3aee15-1546-403a-866a-80455c2abb05-kube-api-access-xwhk2\") pod \"node-ca-5knkr\" (UID: \"1c3aee15-1546-403a-866a-80455c2abb05\") " pod="openshift-image-registry/node-ca-5knkr" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.259543 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1c3aee15-1546-403a-866a-80455c2abb05-host\") pod \"node-ca-5knkr\" (UID: \"1c3aee15-1546-403a-866a-80455c2abb05\") " pod="openshift-image-registry/node-ca-5knkr" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.259530 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.261209 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1c3aee15-1546-403a-866a-80455c2abb05-serviceca\") pod \"node-ca-5knkr\" (UID: \"1c3aee15-1546-403a-866a-80455c2abb05\") " pod="openshift-image-registry/node-ca-5knkr" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.276032 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.290462 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwhk2\" (UniqueName: \"kubernetes.io/projected/1c3aee15-1546-403a-866a-80455c2abb05-kube-api-access-xwhk2\") pod \"node-ca-5knkr\" (UID: \"1c3aee15-1546-403a-866a-80455c2abb05\") " pod="openshift-image-registry/node-ca-5knkr" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.305023 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.328092 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.348938 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.350470 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.350517 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.350531 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.350552 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.350566 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:59Z","lastTransitionTime":"2026-02-27T16:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.368274 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.390988 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.409339 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.409678 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5knkr" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.425074 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: W0227 16:24:59.428552 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c3aee15_1546_403a_866a_80455c2abb05.slice/crio-5ee45494dd085d21f37fa99b434ed6b3f7b03ba418ec0a9913bbd08283f6584b WatchSource:0}: Error finding container 5ee45494dd085d21f37fa99b434ed6b3f7b03ba418ec0a9913bbd08283f6584b: Status 404 returned error can't find the container with id 5ee45494dd085d21f37fa99b434ed6b3f7b03ba418ec0a9913bbd08283f6584b Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.445695 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.453334 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.453394 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.453411 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.453436 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.453452 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:59Z","lastTransitionTime":"2026-02-27T16:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.467187 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.486822 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.486903 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:24:59 crc kubenswrapper[4814]: E0227 16:24:59.487036 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:24:59 crc kubenswrapper[4814]: E0227 16:24:59.487199 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.487309 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:24:59 crc kubenswrapper[4814]: E0227 16:24:59.487397 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.488052 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.507789 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.535541 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.557129 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.557184 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.557201 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.557225 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.557242 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:59Z","lastTransitionTime":"2026-02-27T16:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.561953 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.580704 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.596136 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:24:59Z is after 2025-08-24T17:21:41Z" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.659775 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.659834 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.659855 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.659880 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.659896 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:59Z","lastTransitionTime":"2026-02-27T16:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.764143 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.764201 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.764218 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.764244 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.764311 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:59Z","lastTransitionTime":"2026-02-27T16:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.867748 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.867808 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.867859 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.867882 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.867900 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:59Z","lastTransitionTime":"2026-02-27T16:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.971632 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.971697 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.971717 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.971744 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:24:59 crc kubenswrapper[4814]: I0227 16:24:59.971764 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:24:59Z","lastTransitionTime":"2026-02-27T16:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.001454 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerStarted","Data":"cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef"} Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.001893 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.002126 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.002289 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.011388 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" event={"ID":"2bf0649e-4683-4705-9006-df485b9b5250","Type":"ContainerStarted","Data":"7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28"} Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.014063 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5knkr" event={"ID":"1c3aee15-1546-403a-866a-80455c2abb05","Type":"ContainerStarted","Data":"9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686"} Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.014124 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5knkr" event={"ID":"1c3aee15-1546-403a-866a-80455c2abb05","Type":"ContainerStarted","Data":"5ee45494dd085d21f37fa99b434ed6b3f7b03ba418ec0a9913bbd08283f6584b"} Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.032409 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.049369 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.053423 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.053570 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.078740 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.078710 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.078820 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.078848 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.078881 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.078914 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:00Z","lastTransitionTime":"2026-02-27T16:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.105406 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.125976 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.141706 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.174287 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.181705 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.181759 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.181777 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.181800 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.181818 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:00Z","lastTransitionTime":"2026-02-27T16:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.195727 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.216510 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.247417 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.264556 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.281861 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.284542 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.284610 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.284635 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.284668 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.284689 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:00Z","lastTransitionTime":"2026-02-27T16:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.301668 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.323189 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.340295 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.369002 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.387234 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.387317 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.387331 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.387350 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.387367 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:00Z","lastTransitionTime":"2026-02-27T16:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.388528 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.404511 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.421588 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.442033 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.472737 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.490562 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.490622 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.490644 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.490670 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.490687 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:00Z","lastTransitionTime":"2026-02-27T16:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.493176 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.515069 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.547238 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.562653 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.583128 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.594472 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.594550 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.594568 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.594594 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.594612 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:00Z","lastTransitionTime":"2026-02-27T16:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.603424 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.620549 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:00Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.697850 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.697944 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.697967 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.697994 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.698011 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:00Z","lastTransitionTime":"2026-02-27T16:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.801811 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.801888 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.801908 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.801937 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.801961 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:00Z","lastTransitionTime":"2026-02-27T16:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.935811 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.935874 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.935891 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.935915 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:00 crc kubenswrapper[4814]: I0227 16:25:00.935932 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:00Z","lastTransitionTime":"2026-02-27T16:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.025385 4814 generic.go:334] "Generic (PLEG): container finished" podID="2bf0649e-4683-4705-9006-df485b9b5250" containerID="7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28" exitCode=0 Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.025474 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" event={"ID":"2bf0649e-4683-4705-9006-df485b9b5250","Type":"ContainerDied","Data":"7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28"} Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.039793 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.039869 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.039888 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.040408 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.040476 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:01Z","lastTransitionTime":"2026-02-27T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.051187 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:01Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.084478 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:01Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.110894 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:01Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.131410 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:01Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.144582 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.144666 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.144690 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.144719 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.144740 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:01Z","lastTransitionTime":"2026-02-27T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.150371 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:01Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.165832 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:01Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.183102 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:01Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.202927 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:01Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.223452 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:01Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.245577 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:01Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.247662 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.247824 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.247921 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.248286 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.248311 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:01Z","lastTransitionTime":"2026-02-27T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.265014 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:01Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.286988 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:01Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.316522 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:01Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.335702 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:01Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.352076 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.352111 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.352128 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.352150 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.352166 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:01Z","lastTransitionTime":"2026-02-27T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.454589 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.454634 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.454645 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.454660 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.454670 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:01Z","lastTransitionTime":"2026-02-27T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.486800 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.486839 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.486832 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:01 crc kubenswrapper[4814]: E0227 16:25:01.487164 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:01 crc kubenswrapper[4814]: E0227 16:25:01.487316 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:01 crc kubenswrapper[4814]: E0227 16:25:01.487473 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.487718 4814 scope.go:117] "RemoveContainer" containerID="bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.557980 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.558055 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.558080 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.558112 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.558135 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:01Z","lastTransitionTime":"2026-02-27T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.661422 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.661485 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.661502 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.661527 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.661545 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:01Z","lastTransitionTime":"2026-02-27T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.765013 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.765055 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.765070 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.765093 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.765110 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:01Z","lastTransitionTime":"2026-02-27T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.868281 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.868339 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.868357 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.868381 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.868399 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:01Z","lastTransitionTime":"2026-02-27T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.971335 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.971456 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.971479 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.971508 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:01 crc kubenswrapper[4814]: I0227 16:25:01.971529 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:01Z","lastTransitionTime":"2026-02-27T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.034705 4814 generic.go:334] "Generic (PLEG): container finished" podID="2bf0649e-4683-4705-9006-df485b9b5250" containerID="43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f" exitCode=0 Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.034763 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" event={"ID":"2bf0649e-4683-4705-9006-df485b9b5250","Type":"ContainerDied","Data":"43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f"} Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.045814 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.050145 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680"} Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.050629 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.072916 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.075336 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.075556 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.075645 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.075726 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.075809 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:02Z","lastTransitionTime":"2026-02-27T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.098742 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.119640 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.145631 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.161901 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.177014 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.180336 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.180398 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.180423 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.180451 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.180471 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:02Z","lastTransitionTime":"2026-02-27T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.192631 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.210996 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.228081 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.244105 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.258885 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.273656 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.284778 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.284832 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.284847 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.284921 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.284940 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:02Z","lastTransitionTime":"2026-02-27T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.287589 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.308205 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.329318 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.350212 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.365160 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.382672 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.387659 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.387690 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.387703 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.387720 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.387733 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:02Z","lastTransitionTime":"2026-02-27T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.397848 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.412384 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.436752 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.455362 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.472366 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.491031 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.491241 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.491290 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.491301 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.491316 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.491332 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:02Z","lastTransitionTime":"2026-02-27T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.503985 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.514013 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.538031 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.560238 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:02Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.594350 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.594400 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.594409 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.594424 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.594434 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:02Z","lastTransitionTime":"2026-02-27T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.697873 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.697930 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.697946 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.697970 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.697988 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:02Z","lastTransitionTime":"2026-02-27T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.802022 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.802067 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.802119 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.802144 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.802161 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:02Z","lastTransitionTime":"2026-02-27T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.904613 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.904640 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.904648 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.904660 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:02 crc kubenswrapper[4814]: I0227 16:25:02.904668 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:02Z","lastTransitionTime":"2026-02-27T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.006634 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.006667 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.006677 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.006691 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.006700 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:03Z","lastTransitionTime":"2026-02-27T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.055916 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" event={"ID":"2bf0649e-4683-4705-9006-df485b9b5250","Type":"ContainerStarted","Data":"f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73"} Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.079533 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:03Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.097471 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:03Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.109327 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.109364 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.109375 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.109391 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.109401 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:03Z","lastTransitionTime":"2026-02-27T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.114284 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:03Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.134915 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:03Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.144387 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:03Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.160031 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:03Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.172973 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:03Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.186532 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:03Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.210060 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:03Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.217965 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.218007 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.218018 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.218107 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.218123 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:03Z","lastTransitionTime":"2026-02-27T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.238355 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:03Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.259152 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:03Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.278177 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:03Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.295364 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:03Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.315588 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:03Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.321685 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.321728 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.321739 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.321754 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.321765 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:03Z","lastTransitionTime":"2026-02-27T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.424169 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.424199 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.424208 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.424220 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.424228 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:03Z","lastTransitionTime":"2026-02-27T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.487196 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.487335 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:03 crc kubenswrapper[4814]: E0227 16:25:03.487388 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:03 crc kubenswrapper[4814]: E0227 16:25:03.487514 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.487637 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:03 crc kubenswrapper[4814]: E0227 16:25:03.487712 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.527319 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.527379 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.527397 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.527427 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.527445 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:03Z","lastTransitionTime":"2026-02-27T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.630887 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.630938 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.630948 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.630964 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.630975 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:03Z","lastTransitionTime":"2026-02-27T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.734827 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.734886 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.734906 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.734932 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.734949 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:03Z","lastTransitionTime":"2026-02-27T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.838427 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.838502 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.838522 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.838550 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.838570 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:03Z","lastTransitionTime":"2026-02-27T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.941890 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.941949 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.941964 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.941985 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:03 crc kubenswrapper[4814]: I0227 16:25:03.942000 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:03Z","lastTransitionTime":"2026-02-27T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.047185 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.047359 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.047381 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.047409 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.047428 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:04Z","lastTransitionTime":"2026-02-27T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.064275 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovnkube-controller/0.log" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.070001 4814 generic.go:334] "Generic (PLEG): container finished" podID="4619e86c-7eae-4849-89b0-abb44778d871" containerID="cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef" exitCode=1 Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.070055 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerDied","Data":"cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef"} Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.070911 4814 scope.go:117] "RemoveContainer" containerID="cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.093369 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:04Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.112005 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:04Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.128511 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:04Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.144645 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:04Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.150140 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.150190 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.150213 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.150243 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.150306 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:04Z","lastTransitionTime":"2026-02-27T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.157947 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:04Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.181048 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:04Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.205324 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:04Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.223559 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:04Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.239959 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:04Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.253633 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.253689 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.253708 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.253734 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.253752 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:04Z","lastTransitionTime":"2026-02-27T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.269865 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"message\\\":\\\"2 6597 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 16:25:03.430408 6597 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 16:25:03.430416 6597 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0227 16:25:03.430305 6597 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 16:25:03.430499 6597 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:03.430191 6597 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 16:25:03.430829 6597 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 16:25:03.430845 6597 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:03.431023 6597 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 16:25:03.431467 6597 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0227 16:25:03.431483 6597 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0227 16:25:03.431498 6597 factory.go:656] Stopping watch factory\\\\nI0227 16:25:03.431530 6597 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:03.431536 6597 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:04Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.287427 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:04Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.309803 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:04Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.328453 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:04Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.347911 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:04Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.357069 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.357131 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.357150 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.357175 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.357193 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:04Z","lastTransitionTime":"2026-02-27T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.460383 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.460446 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.460464 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.460491 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.460509 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:04Z","lastTransitionTime":"2026-02-27T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.563294 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.563342 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.563355 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.563372 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.563386 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:04Z","lastTransitionTime":"2026-02-27T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.665579 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.665643 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.665664 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.665689 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.665707 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:04Z","lastTransitionTime":"2026-02-27T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.769107 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.769184 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.769208 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.769237 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.769310 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:04Z","lastTransitionTime":"2026-02-27T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.872550 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.872602 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.872618 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.872643 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.872664 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:04Z","lastTransitionTime":"2026-02-27T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.975758 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.975823 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.975834 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.975856 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:04 crc kubenswrapper[4814]: I0227 16:25:04.975869 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:04Z","lastTransitionTime":"2026-02-27T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.077704 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovnkube-controller/0.log" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.077966 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.078025 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.078038 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.078058 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.078068 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:05Z","lastTransitionTime":"2026-02-27T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.082851 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerStarted","Data":"aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f"} Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.182545 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.182625 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.182647 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.182674 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.182693 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:05Z","lastTransitionTime":"2026-02-27T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.200062 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86"] Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.201356 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.205319 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.205470 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.226734 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.245757 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.260349 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.281857 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.287699 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.287773 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.287795 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.287831 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.287855 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:05Z","lastTransitionTime":"2026-02-27T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.302843 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.326013 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.330761 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.330979 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.331053 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.331106 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.331161 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.331230 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/444e44da-4a93-4115-9be8-7a5a83db72a7-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-h2s86\" (UID: \"444e44da-4a93-4115-9be8-7a5a83db72a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.331331 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.331363 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.331382 4814 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.331431 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 16:25:37.331414597 +0000 UTC m=+149.784039437 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.331340 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/444e44da-4a93-4115-9be8-7a5a83db72a7-env-overrides\") pod \"ovnkube-control-plane-749d76644c-h2s86\" (UID: \"444e44da-4a93-4115-9be8-7a5a83db72a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.331512 4814 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.331556 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trffm\" (UniqueName: \"kubernetes.io/projected/444e44da-4a93-4115-9be8-7a5a83db72a7-kube-api-access-trffm\") pod \"ovnkube-control-plane-749d76644c-h2s86\" (UID: \"444e44da-4a93-4115-9be8-7a5a83db72a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.331644 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/444e44da-4a93-4115-9be8-7a5a83db72a7-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-h2s86\" (UID: \"444e44da-4a93-4115-9be8-7a5a83db72a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.331527 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.331731 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.331743 4814 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.331758 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:25:37.331694737 +0000 UTC m=+149.784319607 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.331556 4814 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.331830 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 16:25:37.33178542 +0000 UTC m=+149.784410290 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.331920 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:25:37.331892284 +0000 UTC m=+149.784517154 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.347075 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.364827 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.387675 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.390773 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.390826 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.390844 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.390872 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.390890 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:05Z","lastTransitionTime":"2026-02-27T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.421278 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.433315 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/444e44da-4a93-4115-9be8-7a5a83db72a7-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-h2s86\" (UID: \"444e44da-4a93-4115-9be8-7a5a83db72a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.433360 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/444e44da-4a93-4115-9be8-7a5a83db72a7-env-overrides\") pod \"ovnkube-control-plane-749d76644c-h2s86\" (UID: \"444e44da-4a93-4115-9be8-7a5a83db72a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.433383 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trffm\" (UniqueName: \"kubernetes.io/projected/444e44da-4a93-4115-9be8-7a5a83db72a7-kube-api-access-trffm\") pod \"ovnkube-control-plane-749d76644c-h2s86\" (UID: \"444e44da-4a93-4115-9be8-7a5a83db72a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.433408 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/444e44da-4a93-4115-9be8-7a5a83db72a7-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-h2s86\" (UID: \"444e44da-4a93-4115-9be8-7a5a83db72a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.434240 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/444e44da-4a93-4115-9be8-7a5a83db72a7-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-h2s86\" (UID: \"444e44da-4a93-4115-9be8-7a5a83db72a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.434820 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/444e44da-4a93-4115-9be8-7a5a83db72a7-env-overrides\") pod \"ovnkube-control-plane-749d76644c-h2s86\" (UID: \"444e44da-4a93-4115-9be8-7a5a83db72a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.440562 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.442691 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/444e44da-4a93-4115-9be8-7a5a83db72a7-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-h2s86\" (UID: \"444e44da-4a93-4115-9be8-7a5a83db72a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.452511 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trffm\" (UniqueName: \"kubernetes.io/projected/444e44da-4a93-4115-9be8-7a5a83db72a7-kube-api-access-trffm\") pod \"ovnkube-control-plane-749d76644c-h2s86\" (UID: \"444e44da-4a93-4115-9be8-7a5a83db72a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.460632 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.481866 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.487125 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.487196 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.487335 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.487428 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.487490 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.487708 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.493809 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.493855 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.493868 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.493887 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.493901 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:05Z","lastTransitionTime":"2026-02-27T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.506580 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"message\\\":\\\"2 6597 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 16:25:03.430408 6597 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 16:25:03.430416 6597 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0227 16:25:03.430305 6597 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 16:25:03.430499 6597 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:03.430191 6597 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 16:25:03.430829 6597 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 16:25:03.430845 6597 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:03.431023 6597 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 16:25:03.431467 6597 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0227 16:25:03.431483 6597 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0227 16:25:03.431498 6597 factory.go:656] Stopping watch factory\\\\nI0227 16:25:03.431530 6597 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:03.431536 6597 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.518688 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.520558 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:05 crc kubenswrapper[4814]: W0227 16:25:05.539535 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod444e44da_4a93_4115_9be8_7a5a83db72a7.slice/crio-0194f557c2f28f13be0c7b94b4d711b1cbee48fbf7df276d3fc25e877785ff49 WatchSource:0}: Error finding container 0194f557c2f28f13be0c7b94b4d711b1cbee48fbf7df276d3fc25e877785ff49: Status 404 returned error can't find the container with id 0194f557c2f28f13be0c7b94b4d711b1cbee48fbf7df276d3fc25e877785ff49 Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.597106 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.597850 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.597869 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.597897 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.597918 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:05Z","lastTransitionTime":"2026-02-27T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.680872 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:25:37.680809412 +0000 UTC m=+150.133434282 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.703519 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.703571 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.703586 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.703604 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.703616 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:05Z","lastTransitionTime":"2026-02-27T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.807149 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.807250 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.807322 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.807357 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.807383 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:05Z","lastTransitionTime":"2026-02-27T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.910338 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.910375 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.910383 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.910401 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.910413 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:05Z","lastTransitionTime":"2026-02-27T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.957728 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-b7hx9"] Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.958286 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:05 crc kubenswrapper[4814]: E0227 16:25:05.958359 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.980902 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:05 crc kubenswrapper[4814]: I0227 16:25:05.998843 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:05Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.012545 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.013565 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.013600 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.013610 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.013626 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.013636 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:06Z","lastTransitionTime":"2026-02-27T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.032211 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.040599 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs\") pod \"network-metrics-daemon-b7hx9\" (UID: \"45306fad-bea6-4c07-92ba-a910f8e26626\") " pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.040656 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccs8k\" (UniqueName: \"kubernetes.io/projected/45306fad-bea6-4c07-92ba-a910f8e26626-kube-api-access-ccs8k\") pod \"network-metrics-daemon-b7hx9\" (UID: \"45306fad-bea6-4c07-92ba-a910f8e26626\") " pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.060831 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"message\\\":\\\"2 6597 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 16:25:03.430408 6597 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 16:25:03.430416 6597 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0227 16:25:03.430305 6597 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 16:25:03.430499 6597 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:03.430191 6597 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 16:25:03.430829 6597 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 16:25:03.430845 6597 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:03.431023 6597 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 16:25:03.431467 6597 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0227 16:25:03.431483 6597 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0227 16:25:03.431498 6597 factory.go:656] Stopping watch factory\\\\nI0227 16:25:03.431530 6597 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:03.431536 6597 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.078646 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.091133 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" event={"ID":"444e44da-4a93-4115-9be8-7a5a83db72a7","Type":"ContainerStarted","Data":"e61727677ea1f23be24d9615f544f3e4e32330117a45dcd459de41369662b388"} Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.091217 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.091240 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" event={"ID":"444e44da-4a93-4115-9be8-7a5a83db72a7","Type":"ContainerStarted","Data":"8784be35b3b5c5fb9061a38828d5af89d6ee8b91181760938a1fa9c13f3b685b"} Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.091287 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" event={"ID":"444e44da-4a93-4115-9be8-7a5a83db72a7","Type":"ContainerStarted","Data":"0194f557c2f28f13be0c7b94b4d711b1cbee48fbf7df276d3fc25e877785ff49"} Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.104730 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45306fad-bea6-4c07-92ba-a910f8e26626\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b7hx9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.116426 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.116459 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.116467 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.116483 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.116494 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:06Z","lastTransitionTime":"2026-02-27T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.128292 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.142068 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccs8k\" (UniqueName: \"kubernetes.io/projected/45306fad-bea6-4c07-92ba-a910f8e26626-kube-api-access-ccs8k\") pod \"network-metrics-daemon-b7hx9\" (UID: \"45306fad-bea6-4c07-92ba-a910f8e26626\") " pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.142155 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs\") pod \"network-metrics-daemon-b7hx9\" (UID: \"45306fad-bea6-4c07-92ba-a910f8e26626\") " pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:06 crc kubenswrapper[4814]: E0227 16:25:06.142880 4814 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 16:25:06 crc kubenswrapper[4814]: E0227 16:25:06.143135 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs podName:45306fad-bea6-4c07-92ba-a910f8e26626 nodeName:}" failed. No retries permitted until 2026-02-27 16:25:06.643119569 +0000 UTC m=+119.095744389 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs") pod "network-metrics-daemon-b7hx9" (UID: "45306fad-bea6-4c07-92ba-a910f8e26626") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.145368 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.164720 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.165169 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccs8k\" (UniqueName: \"kubernetes.io/projected/45306fad-bea6-4c07-92ba-a910f8e26626-kube-api-access-ccs8k\") pod \"network-metrics-daemon-b7hx9\" (UID: \"45306fad-bea6-4c07-92ba-a910f8e26626\") " pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.183520 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.197678 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.210452 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.218619 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.218731 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.218790 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.218867 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.218951 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:06Z","lastTransitionTime":"2026-02-27T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.226010 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.237884 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.237954 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.237971 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.238296 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.238337 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:06Z","lastTransitionTime":"2026-02-27T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.240510 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: E0227 16:25:06.249973 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.253367 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.253883 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.253931 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.253944 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.253964 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.253980 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:06Z","lastTransitionTime":"2026-02-27T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:06 crc kubenswrapper[4814]: E0227 16:25:06.266414 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.267646 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.270111 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.270275 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.270360 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.270450 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.270534 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:06Z","lastTransitionTime":"2026-02-27T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.279156 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: E0227 16:25:06.284946 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.289354 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.289509 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.289611 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.289701 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.289786 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:06Z","lastTransitionTime":"2026-02-27T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.296698 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: E0227 16:25:06.303987 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.309033 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.309086 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.309098 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.309118 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.309132 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:06Z","lastTransitionTime":"2026-02-27T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.315611 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: E0227 16:25:06.325045 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: E0227 16:25:06.325179 4814 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.327105 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.327139 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.327148 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.327190 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.327201 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:06Z","lastTransitionTime":"2026-02-27T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.331800 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.344088 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.361175 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.374805 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.387682 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.409690 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.430812 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.430920 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.430942 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.430967 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.430991 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:06Z","lastTransitionTime":"2026-02-27T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.430930 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.454108 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8784be35b3b5c5fb9061a38828d5af89d6ee8b91181760938a1fa9c13f3b685b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e61727677ea1f23be24d9615f544f3e4e32330117a45dcd459de41369662b388\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.470766 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45306fad-bea6-4c07-92ba-a910f8e26626\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b7hx9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.491744 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.514947 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"message\\\":\\\"2 6597 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 16:25:03.430408 6597 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 16:25:03.430416 6597 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0227 16:25:03.430305 6597 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 16:25:03.430499 6597 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:03.430191 6597 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 16:25:03.430829 6597 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 16:25:03.430845 6597 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:03.431023 6597 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 16:25:03.431467 6597 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0227 16:25:03.431483 6597 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0227 16:25:03.431498 6597 factory.go:656] Stopping watch factory\\\\nI0227 16:25:03.431530 6597 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:03.431536 6597 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.531293 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:06Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.533976 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.534092 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.534111 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.534135 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.534151 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:06Z","lastTransitionTime":"2026-02-27T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.636852 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.636914 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.636932 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.636955 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.636972 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:06Z","lastTransitionTime":"2026-02-27T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.648194 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs\") pod \"network-metrics-daemon-b7hx9\" (UID: \"45306fad-bea6-4c07-92ba-a910f8e26626\") " pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:06 crc kubenswrapper[4814]: E0227 16:25:06.648451 4814 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 16:25:06 crc kubenswrapper[4814]: E0227 16:25:06.648558 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs podName:45306fad-bea6-4c07-92ba-a910f8e26626 nodeName:}" failed. No retries permitted until 2026-02-27 16:25:07.648528817 +0000 UTC m=+120.101153687 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs") pod "network-metrics-daemon-b7hx9" (UID: "45306fad-bea6-4c07-92ba-a910f8e26626") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.740541 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.740621 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.740641 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.740668 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.740699 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:06Z","lastTransitionTime":"2026-02-27T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.843948 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.844024 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.844048 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.844075 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.844096 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:06Z","lastTransitionTime":"2026-02-27T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.947748 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.947860 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.947875 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.947893 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:06 crc kubenswrapper[4814]: I0227 16:25:06.947906 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:06Z","lastTransitionTime":"2026-02-27T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.051953 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.052030 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.052048 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.052075 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.052093 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:07Z","lastTransitionTime":"2026-02-27T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.096395 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovnkube-controller/1.log" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.097730 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovnkube-controller/0.log" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.102281 4814 generic.go:334] "Generic (PLEG): container finished" podID="4619e86c-7eae-4849-89b0-abb44778d871" containerID="aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f" exitCode=1 Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.102323 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerDied","Data":"aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f"} Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.102691 4814 scope.go:117] "RemoveContainer" containerID="cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.104673 4814 scope.go:117] "RemoveContainer" containerID="aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f" Feb 27 16:25:07 crc kubenswrapper[4814]: E0227 16:25:07.104883 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.135313 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:07Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.155585 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:07Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.156539 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.156771 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.156920 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.157080 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.157227 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:07Z","lastTransitionTime":"2026-02-27T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.181678 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8784be35b3b5c5fb9061a38828d5af89d6ee8b91181760938a1fa9c13f3b685b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e61727677ea1f23be24d9615f544f3e4e32330117a45dcd459de41369662b388\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:07Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.204693 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:07Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.228648 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7a777eefa4c07e47de766f7bfa5d400ea08b1b953a4de27e47c87e95d032ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"message\\\":\\\"2 6597 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 16:25:03.430408 6597 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 16:25:03.430416 6597 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0227 16:25:03.430305 6597 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 16:25:03.430499 6597 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:03.430191 6597 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 16:25:03.430829 6597 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 16:25:03.430845 6597 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:03.431023 6597 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 16:25:03.431467 6597 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0227 16:25:03.431483 6597 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0227 16:25:03.431498 6597 factory.go:656] Stopping watch factory\\\\nI0227 16:25:03.431530 6597 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:03.431536 6597 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"ector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:06.311143 6843 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0227 16:25:06.311891 6843 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 16:25:06.311966 6843 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0227 16:25:06.311980 6843 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0227 16:25:06.312008 6843 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 16:25:06.312015 6843 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 16:25:06.312030 6843 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 16:25:06.312051 6843 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0227 16:25:06.312063 6843 factory.go:656] Stopping watch factory\\\\nI0227 16:25:06.312072 6843 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 16:25:06.312082 6843 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:06.312085 6843 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 16:25:06.312094 6843 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0227 16:25:06.312102 6843 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:07Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.243541 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:07Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.261054 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.261109 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.261118 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.261136 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.261148 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:07Z","lastTransitionTime":"2026-02-27T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.263021 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45306fad-bea6-4c07-92ba-a910f8e26626\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b7hx9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:07Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.275845 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:07Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.292436 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:07Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.306239 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:07Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.322122 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:07Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.335701 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:07Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.354980 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:07Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.364339 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.364407 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.364422 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.364452 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.364470 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:07Z","lastTransitionTime":"2026-02-27T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.380509 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:07Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.399993 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:07Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.420621 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:07Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.467750 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.468110 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.468306 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.468498 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.468646 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:07Z","lastTransitionTime":"2026-02-27T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.486714 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.486789 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:07 crc kubenswrapper[4814]: E0227 16:25:07.486874 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.487042 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:07 crc kubenswrapper[4814]: E0227 16:25:07.487046 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.487190 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:07 crc kubenswrapper[4814]: E0227 16:25:07.487267 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:07 crc kubenswrapper[4814]: E0227 16:25:07.488058 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.573604 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.573654 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.573665 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.573681 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.573694 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:07Z","lastTransitionTime":"2026-02-27T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.660909 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs\") pod \"network-metrics-daemon-b7hx9\" (UID: \"45306fad-bea6-4c07-92ba-a910f8e26626\") " pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:07 crc kubenswrapper[4814]: E0227 16:25:07.661091 4814 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 16:25:07 crc kubenswrapper[4814]: E0227 16:25:07.661170 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs podName:45306fad-bea6-4c07-92ba-a910f8e26626 nodeName:}" failed. No retries permitted until 2026-02-27 16:25:09.661150513 +0000 UTC m=+122.113775343 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs") pod "network-metrics-daemon-b7hx9" (UID: "45306fad-bea6-4c07-92ba-a910f8e26626") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.677162 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.677212 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.677230 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.677267 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.677280 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:07Z","lastTransitionTime":"2026-02-27T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.780785 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.780850 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.780866 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.780885 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.780899 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:07Z","lastTransitionTime":"2026-02-27T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.883574 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.883648 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.883663 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.883688 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.883709 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:07Z","lastTransitionTime":"2026-02-27T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.987364 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.987422 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.987440 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.987466 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:07 crc kubenswrapper[4814]: I0227 16:25:07.987487 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:07Z","lastTransitionTime":"2026-02-27T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.090053 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.090114 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.090145 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.090176 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.090200 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:08Z","lastTransitionTime":"2026-02-27T16:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.108084 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovnkube-controller/1.log" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.112879 4814 scope.go:117] "RemoveContainer" containerID="aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f" Feb 27 16:25:08 crc kubenswrapper[4814]: E0227 16:25:08.113138 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.132673 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.150716 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.167716 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.188619 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.193308 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.193388 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.193406 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.193432 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.193449 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:08Z","lastTransitionTime":"2026-02-27T16:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.208697 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.228487 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.247871 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.270770 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.290914 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.296610 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.297329 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.297492 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.297634 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.297777 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:08Z","lastTransitionTime":"2026-02-27T16:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.303670 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8784be35b3b5c5fb9061a38828d5af89d6ee8b91181760938a1fa9c13f3b685b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e61727677ea1f23be24d9615f544f3e4e32330117a45dcd459de41369662b388\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.331698 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.345828 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.355400 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.366030 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45306fad-bea6-4c07-92ba-a910f8e26626\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b7hx9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.378544 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.396942 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"ector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:06.311143 6843 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0227 16:25:06.311891 6843 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 16:25:06.311966 6843 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0227 16:25:06.311980 6843 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0227 16:25:06.312008 6843 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 16:25:06.312015 6843 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 16:25:06.312030 6843 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 16:25:06.312051 6843 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0227 16:25:06.312063 6843 factory.go:656] Stopping watch factory\\\\nI0227 16:25:06.312072 6843 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 16:25:06.312082 6843 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:06.312085 6843 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 16:25:06.312094 6843 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0227 16:25:06.312102 6843 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.400404 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.400462 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.400477 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.400494 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.400507 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:08Z","lastTransitionTime":"2026-02-27T16:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:08 crc kubenswrapper[4814]: E0227 16:25:08.501750 4814 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.517107 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"ector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:06.311143 6843 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0227 16:25:06.311891 6843 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 16:25:06.311966 6843 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0227 16:25:06.311980 6843 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0227 16:25:06.312008 6843 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 16:25:06.312015 6843 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 16:25:06.312030 6843 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 16:25:06.312051 6843 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0227 16:25:06.312063 6843 factory.go:656] Stopping watch factory\\\\nI0227 16:25:06.312072 6843 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 16:25:06.312082 6843 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:06.312085 6843 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 16:25:06.312094 6843 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0227 16:25:06.312102 6843 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.539295 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.558063 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45306fad-bea6-4c07-92ba-a910f8e26626\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b7hx9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.575624 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.591839 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: E0227 16:25:08.594595 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.610685 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.622747 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.635831 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.650020 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.666173 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.680166 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.691343 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.709055 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.722949 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.737341 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8784be35b3b5c5fb9061a38828d5af89d6ee8b91181760938a1fa9c13f3b685b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e61727677ea1f23be24d9615f544f3e4e32330117a45dcd459de41369662b388\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:08 crc kubenswrapper[4814]: I0227 16:25:08.768805 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:08Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:09 crc kubenswrapper[4814]: I0227 16:25:09.487606 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:09 crc kubenswrapper[4814]: I0227 16:25:09.487613 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:09 crc kubenswrapper[4814]: I0227 16:25:09.488001 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:09 crc kubenswrapper[4814]: E0227 16:25:09.487844 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:09 crc kubenswrapper[4814]: I0227 16:25:09.487657 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:09 crc kubenswrapper[4814]: E0227 16:25:09.488141 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:09 crc kubenswrapper[4814]: E0227 16:25:09.488239 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:09 crc kubenswrapper[4814]: E0227 16:25:09.488402 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:09 crc kubenswrapper[4814]: I0227 16:25:09.683792 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs\") pod \"network-metrics-daemon-b7hx9\" (UID: \"45306fad-bea6-4c07-92ba-a910f8e26626\") " pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:09 crc kubenswrapper[4814]: E0227 16:25:09.684077 4814 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 16:25:09 crc kubenswrapper[4814]: E0227 16:25:09.684236 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs podName:45306fad-bea6-4c07-92ba-a910f8e26626 nodeName:}" failed. No retries permitted until 2026-02-27 16:25:13.684201219 +0000 UTC m=+126.136826239 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs") pod "network-metrics-daemon-b7hx9" (UID: "45306fad-bea6-4c07-92ba-a910f8e26626") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 16:25:11 crc kubenswrapper[4814]: I0227 16:25:11.486956 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:11 crc kubenswrapper[4814]: I0227 16:25:11.487076 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:11 crc kubenswrapper[4814]: I0227 16:25:11.487175 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:11 crc kubenswrapper[4814]: I0227 16:25:11.487194 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:11 crc kubenswrapper[4814]: E0227 16:25:11.487095 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:11 crc kubenswrapper[4814]: E0227 16:25:11.487356 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:11 crc kubenswrapper[4814]: E0227 16:25:11.487424 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:11 crc kubenswrapper[4814]: E0227 16:25:11.487513 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:13 crc kubenswrapper[4814]: I0227 16:25:13.486837 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:13 crc kubenswrapper[4814]: I0227 16:25:13.486907 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:13 crc kubenswrapper[4814]: I0227 16:25:13.486952 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:13 crc kubenswrapper[4814]: I0227 16:25:13.486863 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:13 crc kubenswrapper[4814]: E0227 16:25:13.487023 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:13 crc kubenswrapper[4814]: E0227 16:25:13.487153 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:13 crc kubenswrapper[4814]: E0227 16:25:13.487230 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:13 crc kubenswrapper[4814]: E0227 16:25:13.487420 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:13 crc kubenswrapper[4814]: E0227 16:25:13.595676 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:25:13 crc kubenswrapper[4814]: I0227 16:25:13.730845 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs\") pod \"network-metrics-daemon-b7hx9\" (UID: \"45306fad-bea6-4c07-92ba-a910f8e26626\") " pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:13 crc kubenswrapper[4814]: E0227 16:25:13.731045 4814 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 16:25:13 crc kubenswrapper[4814]: E0227 16:25:13.731159 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs podName:45306fad-bea6-4c07-92ba-a910f8e26626 nodeName:}" failed. No retries permitted until 2026-02-27 16:25:21.731125262 +0000 UTC m=+134.183750132 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs") pod "network-metrics-daemon-b7hx9" (UID: "45306fad-bea6-4c07-92ba-a910f8e26626") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 16:25:15 crc kubenswrapper[4814]: I0227 16:25:15.487442 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:15 crc kubenswrapper[4814]: I0227 16:25:15.487531 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:15 crc kubenswrapper[4814]: I0227 16:25:15.487465 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:15 crc kubenswrapper[4814]: E0227 16:25:15.487620 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:15 crc kubenswrapper[4814]: E0227 16:25:15.487681 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:15 crc kubenswrapper[4814]: I0227 16:25:15.487489 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:15 crc kubenswrapper[4814]: E0227 16:25:15.487769 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:15 crc kubenswrapper[4814]: E0227 16:25:15.487886 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.501557 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.668462 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.668512 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.668527 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.668551 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.668568 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:16Z","lastTransitionTime":"2026-02-27T16:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:16 crc kubenswrapper[4814]: E0227 16:25:16.691675 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:16Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.696818 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.696865 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.696876 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.696892 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.696902 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:16Z","lastTransitionTime":"2026-02-27T16:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:16 crc kubenswrapper[4814]: E0227 16:25:16.717246 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:16Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.721691 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.721738 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.721750 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.721765 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.721777 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:16Z","lastTransitionTime":"2026-02-27T16:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:16 crc kubenswrapper[4814]: E0227 16:25:16.739814 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:16Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.744629 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.744662 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.744672 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.744684 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.744694 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:16Z","lastTransitionTime":"2026-02-27T16:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:16 crc kubenswrapper[4814]: E0227 16:25:16.759726 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:16Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.764423 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.764479 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.764491 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.764523 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:16 crc kubenswrapper[4814]: I0227 16:25:16.764537 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:16Z","lastTransitionTime":"2026-02-27T16:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:16 crc kubenswrapper[4814]: E0227 16:25:16.785876 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:16Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:16 crc kubenswrapper[4814]: E0227 16:25:16.786064 4814 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 16:25:17 crc kubenswrapper[4814]: I0227 16:25:17.487575 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:17 crc kubenswrapper[4814]: I0227 16:25:17.487634 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:17 crc kubenswrapper[4814]: I0227 16:25:17.487711 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:17 crc kubenswrapper[4814]: I0227 16:25:17.487666 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:17 crc kubenswrapper[4814]: E0227 16:25:17.487824 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:17 crc kubenswrapper[4814]: E0227 16:25:17.487907 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:17 crc kubenswrapper[4814]: E0227 16:25:17.488028 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:17 crc kubenswrapper[4814]: E0227 16:25:17.488155 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.505749 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.512571 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.532869 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.551278 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8784be35b3b5c5fb9061a38828d5af89d6ee8b91181760938a1fa9c13f3b685b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e61727677ea1f23be24d9615f544f3e4e32330117a45dcd459de41369662b388\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.568073 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45306fad-bea6-4c07-92ba-a910f8e26626\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b7hx9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.589475 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:18 crc kubenswrapper[4814]: E0227 16:25:18.596417 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.612890 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"ector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:06.311143 6843 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0227 16:25:06.311891 6843 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 16:25:06.311966 6843 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0227 16:25:06.311980 6843 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0227 16:25:06.312008 6843 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 16:25:06.312015 6843 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 16:25:06.312030 6843 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 16:25:06.312051 6843 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0227 16:25:06.312063 6843 factory.go:656] Stopping watch factory\\\\nI0227 16:25:06.312072 6843 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 16:25:06.312082 6843 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:06.312085 6843 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 16:25:06.312094 6843 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0227 16:25:06.312102 6843 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.627776 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.642208 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.655651 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.677343 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.697024 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.717116 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.734387 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.754154 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.773491 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36413f7d-bc90-4ab1-8cbd-e9a7a36f92cf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe23574a71dd13340a8be7b9dae77410b55fd1daa11a58eb8ed914fb44cea829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34064bf25c31b4fb36554561a8cf9109191e950be41e22c71e9d4773e0e66ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc29a3f6d746e0eda7be7e511df92aa978993ac600a788820e292603b328434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.790549 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:18 crc kubenswrapper[4814]: I0227 16:25:18.810474 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:18Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:19 crc kubenswrapper[4814]: I0227 16:25:19.486854 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:19 crc kubenswrapper[4814]: I0227 16:25:19.486969 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:19 crc kubenswrapper[4814]: E0227 16:25:19.487037 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:19 crc kubenswrapper[4814]: I0227 16:25:19.487036 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:19 crc kubenswrapper[4814]: I0227 16:25:19.487129 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:19 crc kubenswrapper[4814]: E0227 16:25:19.487351 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:19 crc kubenswrapper[4814]: E0227 16:25:19.487547 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:19 crc kubenswrapper[4814]: E0227 16:25:19.487677 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.284454 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.307757 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8269fdbb-d61c-4b25-8fb1-251ee6857851\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22d8635e425d164cd7c63685f0fb8b64a69919a2a687c9ca42321bfed7559ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5ba138e1d2cbeab8ba50d4d02b246c31cb1226615131fa4b074a08d615ca4f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:23:41Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 16:23:10.695873 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 16:23:10.702271 1 observer_polling.go:159] Starting file observer\\\\nI0227 16:23:10.732327 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 16:23:10.742605 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0227 16:23:41.467207 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:40Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501ea405666a49080dd2af178db81a992757c23da06fedd1a6036047880f3f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbf88717d3b766e6b99c0ec623769bd67642fe95222e221d59f15e0e2fdf3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86c2da64ed7a7f6af14d9fdd54dbb1ace65f683dc22f739afbffa9f1cd6d52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.329480 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.344837 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.360874 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.381325 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36413f7d-bc90-4ab1-8cbd-e9a7a36f92cf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe23574a71dd13340a8be7b9dae77410b55fd1daa11a58eb8ed914fb44cea829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34064bf25c31b4fb36554561a8cf9109191e950be41e22c71e9d4773e0e66ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc29a3f6d746e0eda7be7e511df92aa978993ac600a788820e292603b328434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.401102 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.422294 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.441395 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.461469 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.475969 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.502027 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.538816 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.560946 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.581313 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8784be35b3b5c5fb9061a38828d5af89d6ee8b91181760938a1fa9c13f3b685b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e61727677ea1f23be24d9615f544f3e4e32330117a45dcd459de41369662b388\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.602672 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.628312 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"ector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:06.311143 6843 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0227 16:25:06.311891 6843 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 16:25:06.311966 6843 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0227 16:25:06.311980 6843 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0227 16:25:06.312008 6843 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 16:25:06.312015 6843 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 16:25:06.312030 6843 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 16:25:06.312051 6843 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0227 16:25:06.312063 6843 factory.go:656] Stopping watch factory\\\\nI0227 16:25:06.312072 6843 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 16:25:06.312082 6843 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:06.312085 6843 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 16:25:06.312094 6843 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0227 16:25:06.312102 6843 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.647751 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:20 crc kubenswrapper[4814]: I0227 16:25:20.664142 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45306fad-bea6-4c07-92ba-a910f8e26626\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b7hx9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:20Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:21 crc kubenswrapper[4814]: I0227 16:25:21.486684 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:21 crc kubenswrapper[4814]: I0227 16:25:21.486685 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:21 crc kubenswrapper[4814]: I0227 16:25:21.486715 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:21 crc kubenswrapper[4814]: I0227 16:25:21.486878 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:21 crc kubenswrapper[4814]: E0227 16:25:21.487090 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:21 crc kubenswrapper[4814]: E0227 16:25:21.487273 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:21 crc kubenswrapper[4814]: E0227 16:25:21.487462 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:21 crc kubenswrapper[4814]: E0227 16:25:21.487567 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:21 crc kubenswrapper[4814]: I0227 16:25:21.831385 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs\") pod \"network-metrics-daemon-b7hx9\" (UID: \"45306fad-bea6-4c07-92ba-a910f8e26626\") " pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:21 crc kubenswrapper[4814]: E0227 16:25:21.831588 4814 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 16:25:21 crc kubenswrapper[4814]: E0227 16:25:21.831671 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs podName:45306fad-bea6-4c07-92ba-a910f8e26626 nodeName:}" failed. No retries permitted until 2026-02-27 16:25:37.831649824 +0000 UTC m=+150.284274684 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs") pod "network-metrics-daemon-b7hx9" (UID: "45306fad-bea6-4c07-92ba-a910f8e26626") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 16:25:22 crc kubenswrapper[4814]: I0227 16:25:22.489201 4814 scope.go:117] "RemoveContainer" containerID="aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.178782 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovnkube-controller/1.log" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.182739 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerStarted","Data":"4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990"} Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.183107 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.199418 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.227387 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.241066 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.257999 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.270473 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36413f7d-bc90-4ab1-8cbd-e9a7a36f92cf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe23574a71dd13340a8be7b9dae77410b55fd1daa11a58eb8ed914fb44cea829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34064bf25c31b4fb36554561a8cf9109191e950be41e22c71e9d4773e0e66ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc29a3f6d746e0eda7be7e511df92aa978993ac600a788820e292603b328434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.282131 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.295959 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.318321 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.332397 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.342505 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8784be35b3b5c5fb9061a38828d5af89d6ee8b91181760938a1fa9c13f3b685b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e61727677ea1f23be24d9615f544f3e4e32330117a45dcd459de41369662b388\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.352377 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45306fad-bea6-4c07-92ba-a910f8e26626\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b7hx9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.368781 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.393887 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"ector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:06.311143 6843 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0227 16:25:06.311891 6843 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 16:25:06.311966 6843 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0227 16:25:06.311980 6843 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0227 16:25:06.312008 6843 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 16:25:06.312015 6843 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 16:25:06.312030 6843 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 16:25:06.312051 6843 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0227 16:25:06.312063 6843 factory.go:656] Stopping watch factory\\\\nI0227 16:25:06.312072 6843 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 16:25:06.312082 6843 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:06.312085 6843 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 16:25:06.312094 6843 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0227 16:25:06.312102 6843 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.407349 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.428627 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.441346 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.458452 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8269fdbb-d61c-4b25-8fb1-251ee6857851\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22d8635e425d164cd7c63685f0fb8b64a69919a2a687c9ca42321bfed7559ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5ba138e1d2cbeab8ba50d4d02b246c31cb1226615131fa4b074a08d615ca4f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:23:41Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 16:23:10.695873 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 16:23:10.702271 1 observer_polling.go:159] Starting file observer\\\\nI0227 16:23:10.732327 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 16:23:10.742605 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0227 16:23:41.467207 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:40Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501ea405666a49080dd2af178db81a992757c23da06fedd1a6036047880f3f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbf88717d3b766e6b99c0ec623769bd67642fe95222e221d59f15e0e2fdf3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86c2da64ed7a7f6af14d9fdd54dbb1ace65f683dc22f739afbffa9f1cd6d52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.473688 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:23Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.487426 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:23 crc kubenswrapper[4814]: E0227 16:25:23.487586 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.487675 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:23 crc kubenswrapper[4814]: E0227 16:25:23.487735 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.487788 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:23 crc kubenswrapper[4814]: E0227 16:25:23.487847 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:23 crc kubenswrapper[4814]: I0227 16:25:23.487899 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:23 crc kubenswrapper[4814]: E0227 16:25:23.488002 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:23 crc kubenswrapper[4814]: E0227 16:25:23.597694 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.188801 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovnkube-controller/2.log" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.189753 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovnkube-controller/1.log" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.194024 4814 generic.go:334] "Generic (PLEG): container finished" podID="4619e86c-7eae-4849-89b0-abb44778d871" containerID="4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990" exitCode=1 Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.194086 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerDied","Data":"4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990"} Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.194133 4814 scope.go:117] "RemoveContainer" containerID="aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.195313 4814 scope.go:117] "RemoveContainer" containerID="4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990" Feb 27 16:25:24 crc kubenswrapper[4814]: E0227 16:25:24.195903 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.217778 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa1fc6171afc2842696a5af951cb754cdb61c4533ff87b392144b640e34e489f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"message\\\":\\\"ector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 16:25:06.311143 6843 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0227 16:25:06.311891 6843 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 16:25:06.311966 6843 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0227 16:25:06.311980 6843 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0227 16:25:06.312008 6843 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 16:25:06.312015 6843 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 16:25:06.312030 6843 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 16:25:06.312051 6843 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0227 16:25:06.312063 6843 factory.go:656] Stopping watch factory\\\\nI0227 16:25:06.312072 6843 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0227 16:25:06.312082 6843 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:06.312085 6843 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 16:25:06.312094 6843 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0227 16:25:06.312102 6843 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:23Z\\\",\\\"message\\\":\\\"/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0227 16:25:23.572210 7108 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 16:25:23.572245 7108 factory.go:656] Stopping watch factory\\\\nI0227 16:25:23.572272 7108 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:23.572302 7108 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 16:25:23.572311 7108 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 16:25:23.572406 7108 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.230296 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.241809 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45306fad-bea6-4c07-92ba-a910f8e26626\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b7hx9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.254985 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.272859 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8269fdbb-d61c-4b25-8fb1-251ee6857851\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22d8635e425d164cd7c63685f0fb8b64a69919a2a687c9ca42321bfed7559ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5ba138e1d2cbeab8ba50d4d02b246c31cb1226615131fa4b074a08d615ca4f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:23:41Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 16:23:10.695873 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 16:23:10.702271 1 observer_polling.go:159] Starting file observer\\\\nI0227 16:23:10.732327 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 16:23:10.742605 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0227 16:23:41.467207 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:40Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501ea405666a49080dd2af178db81a992757c23da06fedd1a6036047880f3f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbf88717d3b766e6b99c0ec623769bd67642fe95222e221d59f15e0e2fdf3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86c2da64ed7a7f6af14d9fdd54dbb1ace65f683dc22f739afbffa9f1cd6d52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.293390 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.306322 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.319234 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.330817 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.347461 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.363721 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.382534 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.400681 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.423120 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.438922 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36413f7d-bc90-4ab1-8cbd-e9a7a36f92cf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe23574a71dd13340a8be7b9dae77410b55fd1daa11a58eb8ed914fb44cea829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34064bf25c31b4fb36554561a8cf9109191e950be41e22c71e9d4773e0e66ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc29a3f6d746e0eda7be7e511df92aa978993ac600a788820e292603b328434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.457866 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.475406 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8784be35b3b5c5fb9061a38828d5af89d6ee8b91181760938a1fa9c13f3b685b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e61727677ea1f23be24d9615f544f3e4e32330117a45dcd459de41369662b388\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:24 crc kubenswrapper[4814]: I0227 16:25:24.499451 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:24Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.200511 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovnkube-controller/2.log" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.204211 4814 scope.go:117] "RemoveContainer" containerID="4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990" Feb 27 16:25:25 crc kubenswrapper[4814]: E0227 16:25:25.204385 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.220298 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.238628 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8269fdbb-d61c-4b25-8fb1-251ee6857851\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22d8635e425d164cd7c63685f0fb8b64a69919a2a687c9ca42321bfed7559ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5ba138e1d2cbeab8ba50d4d02b246c31cb1226615131fa4b074a08d615ca4f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:23:41Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 16:23:10.695873 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 16:23:10.702271 1 observer_polling.go:159] Starting file observer\\\\nI0227 16:23:10.732327 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 16:23:10.742605 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0227 16:23:41.467207 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:40Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501ea405666a49080dd2af178db81a992757c23da06fedd1a6036047880f3f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbf88717d3b766e6b99c0ec623769bd67642fe95222e221d59f15e0e2fdf3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86c2da64ed7a7f6af14d9fdd54dbb1ace65f683dc22f739afbffa9f1cd6d52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.261440 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.280072 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.295455 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.308412 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.330459 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.349574 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36413f7d-bc90-4ab1-8cbd-e9a7a36f92cf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe23574a71dd13340a8be7b9dae77410b55fd1daa11a58eb8ed914fb44cea829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34064bf25c31b4fb36554561a8cf9109191e950be41e22c71e9d4773e0e66ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc29a3f6d746e0eda7be7e511df92aa978993ac600a788820e292603b328434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.369275 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.387477 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.405908 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.431492 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.450871 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.467574 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8784be35b3b5c5fb9061a38828d5af89d6ee8b91181760938a1fa9c13f3b685b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e61727677ea1f23be24d9615f544f3e4e32330117a45dcd459de41369662b388\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.485846 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.487018 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.487171 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.487316 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.487361 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:25 crc kubenswrapper[4814]: E0227 16:25:25.487327 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:25 crc kubenswrapper[4814]: E0227 16:25:25.487445 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:25 crc kubenswrapper[4814]: E0227 16:25:25.487621 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:25 crc kubenswrapper[4814]: E0227 16:25:25.487705 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.509125 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:23Z\\\",\\\"message\\\":\\\"/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0227 16:25:23.572210 7108 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 16:25:23.572245 7108 factory.go:656] Stopping watch factory\\\\nI0227 16:25:23.572272 7108 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:23.572302 7108 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 16:25:23.572311 7108 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 16:25:23.572406 7108 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.520684 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:25 crc kubenswrapper[4814]: I0227 16:25:25.535030 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45306fad-bea6-4c07-92ba-a910f8e26626\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b7hx9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:25Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:26 crc kubenswrapper[4814]: I0227 16:25:26.979547 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:26 crc kubenswrapper[4814]: I0227 16:25:26.979630 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:26 crc kubenswrapper[4814]: I0227 16:25:26.979649 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:26 crc kubenswrapper[4814]: I0227 16:25:26.979672 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:26 crc kubenswrapper[4814]: I0227 16:25:26.979693 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:26Z","lastTransitionTime":"2026-02-27T16:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:26 crc kubenswrapper[4814]: E0227 16:25:26.998451 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:26Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.003282 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.003324 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.003337 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.003359 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.003373 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:27Z","lastTransitionTime":"2026-02-27T16:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:27 crc kubenswrapper[4814]: E0227 16:25:27.019059 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:27Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.024973 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.025013 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.025025 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.025043 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.025056 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:27Z","lastTransitionTime":"2026-02-27T16:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:27 crc kubenswrapper[4814]: E0227 16:25:27.046557 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:27Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.051820 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.051880 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.051900 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.051926 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.051943 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:27Z","lastTransitionTime":"2026-02-27T16:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:27 crc kubenswrapper[4814]: E0227 16:25:27.070944 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:27Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.076390 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.076443 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.076461 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.076487 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.076508 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:27Z","lastTransitionTime":"2026-02-27T16:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:27 crc kubenswrapper[4814]: E0227 16:25:27.097377 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:27Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:27 crc kubenswrapper[4814]: E0227 16:25:27.097606 4814 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.487517 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.487600 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.487600 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:27 crc kubenswrapper[4814]: E0227 16:25:27.488240 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:27 crc kubenswrapper[4814]: I0227 16:25:27.487726 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:27 crc kubenswrapper[4814]: E0227 16:25:27.488423 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:27 crc kubenswrapper[4814]: E0227 16:25:27.488471 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:27 crc kubenswrapper[4814]: E0227 16:25:27.488070 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.507679 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.518509 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8784be35b3b5c5fb9061a38828d5af89d6ee8b91181760938a1fa9c13f3b685b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e61727677ea1f23be24d9615f544f3e4e32330117a45dcd459de41369662b388\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.540784 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.565417 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:23Z\\\",\\\"message\\\":\\\"/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0227 16:25:23.572210 7108 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 16:25:23.572245 7108 factory.go:656] Stopping watch factory\\\\nI0227 16:25:23.572272 7108 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:23.572302 7108 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 16:25:23.572311 7108 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 16:25:23.572406 7108 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.581576 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.595652 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45306fad-bea6-4c07-92ba-a910f8e26626\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b7hx9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: E0227 16:25:28.598819 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.615425 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.632116 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8269fdbb-d61c-4b25-8fb1-251ee6857851\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22d8635e425d164cd7c63685f0fb8b64a69919a2a687c9ca42321bfed7559ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5ba138e1d2cbeab8ba50d4d02b246c31cb1226615131fa4b074a08d615ca4f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:23:41Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 16:23:10.695873 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 16:23:10.702271 1 observer_polling.go:159] Starting file observer\\\\nI0227 16:23:10.732327 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 16:23:10.742605 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0227 16:23:41.467207 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:40Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501ea405666a49080dd2af178db81a992757c23da06fedd1a6036047880f3f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbf88717d3b766e6b99c0ec623769bd67642fe95222e221d59f15e0e2fdf3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86c2da64ed7a7f6af14d9fdd54dbb1ace65f683dc22f739afbffa9f1cd6d52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.648157 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.662359 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.682033 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.698734 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.713676 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.729633 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.747981 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.763495 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.781328 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:28 crc kubenswrapper[4814]: I0227 16:25:28.796971 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36413f7d-bc90-4ab1-8cbd-e9a7a36f92cf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe23574a71dd13340a8be7b9dae77410b55fd1daa11a58eb8ed914fb44cea829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34064bf25c31b4fb36554561a8cf9109191e950be41e22c71e9d4773e0e66ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc29a3f6d746e0eda7be7e511df92aa978993ac600a788820e292603b328434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:28Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:29 crc kubenswrapper[4814]: I0227 16:25:29.486973 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:29 crc kubenswrapper[4814]: I0227 16:25:29.486998 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:29 crc kubenswrapper[4814]: I0227 16:25:29.487067 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:29 crc kubenswrapper[4814]: I0227 16:25:29.487293 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:29 crc kubenswrapper[4814]: E0227 16:25:29.487327 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:29 crc kubenswrapper[4814]: E0227 16:25:29.487389 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:29 crc kubenswrapper[4814]: E0227 16:25:29.487511 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:29 crc kubenswrapper[4814]: E0227 16:25:29.487643 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:31 crc kubenswrapper[4814]: I0227 16:25:31.487678 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:31 crc kubenswrapper[4814]: I0227 16:25:31.487731 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:31 crc kubenswrapper[4814]: I0227 16:25:31.487805 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:31 crc kubenswrapper[4814]: E0227 16:25:31.487966 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:31 crc kubenswrapper[4814]: I0227 16:25:31.487990 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:31 crc kubenswrapper[4814]: E0227 16:25:31.488137 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:31 crc kubenswrapper[4814]: E0227 16:25:31.488413 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:31 crc kubenswrapper[4814]: E0227 16:25:31.488587 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:33 crc kubenswrapper[4814]: I0227 16:25:33.486970 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:33 crc kubenswrapper[4814]: I0227 16:25:33.487139 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:33 crc kubenswrapper[4814]: E0227 16:25:33.487479 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:33 crc kubenswrapper[4814]: I0227 16:25:33.487943 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:33 crc kubenswrapper[4814]: I0227 16:25:33.488016 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:33 crc kubenswrapper[4814]: E0227 16:25:33.488118 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:33 crc kubenswrapper[4814]: E0227 16:25:33.488403 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:33 crc kubenswrapper[4814]: E0227 16:25:33.488635 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:33 crc kubenswrapper[4814]: E0227 16:25:33.600384 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:25:35 crc kubenswrapper[4814]: I0227 16:25:35.487401 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:35 crc kubenswrapper[4814]: I0227 16:25:35.487468 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:35 crc kubenswrapper[4814]: I0227 16:25:35.487428 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:35 crc kubenswrapper[4814]: I0227 16:25:35.487418 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:35 crc kubenswrapper[4814]: E0227 16:25:35.487679 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:35 crc kubenswrapper[4814]: E0227 16:25:35.487816 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:35 crc kubenswrapper[4814]: E0227 16:25:35.488000 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:35 crc kubenswrapper[4814]: E0227 16:25:35.488059 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.171758 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.171839 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.171858 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.171901 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.171923 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:37Z","lastTransitionTime":"2026-02-27T16:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.195718 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:37Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.201198 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.201299 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.201327 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.201362 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.201386 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:37Z","lastTransitionTime":"2026-02-27T16:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.222230 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:37Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.232945 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.233066 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.233087 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.233249 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.233406 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:37Z","lastTransitionTime":"2026-02-27T16:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.256377 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:37Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.261983 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.262028 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.262044 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.262063 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.262076 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:37Z","lastTransitionTime":"2026-02-27T16:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.278511 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:37Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.285019 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.285092 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.285108 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.285134 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.285150 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:37Z","lastTransitionTime":"2026-02-27T16:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.304459 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:37Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.304687 4814 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.406130 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.406232 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.406323 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.406405 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.406589 4814 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.406640 4814 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.406648 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.406723 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:26:41.406690801 +0000 UTC m=+213.859315801 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.406709 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.406754 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.406766 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.406778 4814 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.406792 4814 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.406746 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:26:41.406737633 +0000 UTC m=+213.859362673 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.406882 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 16:26:41.406852847 +0000 UTC m=+213.859477707 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.406911 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 16:26:41.406898969 +0000 UTC m=+213.859523849 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.487649 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.487787 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.487815 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.488073 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.488122 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.488482 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.488537 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.488607 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.506427 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.709975 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.710157 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:41.710122609 +0000 UTC m=+214.162747459 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:25:37 crc kubenswrapper[4814]: I0227 16:25:37.912974 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs\") pod \"network-metrics-daemon-b7hx9\" (UID: \"45306fad-bea6-4c07-92ba-a910f8e26626\") " pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.913285 4814 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 16:25:37 crc kubenswrapper[4814]: E0227 16:25:37.913419 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs podName:45306fad-bea6-4c07-92ba-a910f8e26626 nodeName:}" failed. No retries permitted until 2026-02-27 16:26:09.913383937 +0000 UTC m=+182.366008807 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs") pod "network-metrics-daemon-b7hx9" (UID: "45306fad-bea6-4c07-92ba-a910f8e26626") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.488228 4814 scope.go:117] "RemoveContainer" containerID="4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990" Feb 27 16:25:38 crc kubenswrapper[4814]: E0227 16:25:38.488610 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.508061 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36413f7d-bc90-4ab1-8cbd-e9a7a36f92cf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe23574a71dd13340a8be7b9dae77410b55fd1daa11a58eb8ed914fb44cea829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34064bf25c31b4fb36554561a8cf9109191e950be41e22c71e9d4773e0e66ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc29a3f6d746e0eda7be7e511df92aa978993ac600a788820e292603b328434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.531472 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.550883 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.570237 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.590099 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: E0227 16:25:38.610125 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.641666 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.669935 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.701172 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.720126 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.736372 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8784be35b3b5c5fb9061a38828d5af89d6ee8b91181760938a1fa9c13f3b685b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e61727677ea1f23be24d9615f544f3e4e32330117a45dcd459de41369662b388\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.753160 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.780151 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:23Z\\\",\\\"message\\\":\\\"/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0227 16:25:23.572210 7108 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 16:25:23.572245 7108 factory.go:656] Stopping watch factory\\\\nI0227 16:25:23.572272 7108 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:23.572302 7108 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 16:25:23.572311 7108 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 16:25:23.572406 7108 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.797798 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.812911 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45306fad-bea6-4c07-92ba-a910f8e26626\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b7hx9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.828201 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d657f969-c1a7-4c81-ad14-57a0d0660d46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d1c14a07903094a68c4b62d08ede23259920fb342412bc54a5d1ce0ea92b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14102163d15d2563d01aa8d820f93aa6c36a244adb644431399aa94fea3b568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14102163d15d2563d01aa8d820f93aa6c36a244adb644431399aa94fea3b568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.846608 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8269fdbb-d61c-4b25-8fb1-251ee6857851\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22d8635e425d164cd7c63685f0fb8b64a69919a2a687c9ca42321bfed7559ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5ba138e1d2cbeab8ba50d4d02b246c31cb1226615131fa4b074a08d615ca4f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:23:41Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 16:23:10.695873 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 16:23:10.702271 1 observer_polling.go:159] Starting file observer\\\\nI0227 16:23:10.732327 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 16:23:10.742605 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0227 16:23:41.467207 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:40Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501ea405666a49080dd2af178db81a992757c23da06fedd1a6036047880f3f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbf88717d3b766e6b99c0ec623769bd67642fe95222e221d59f15e0e2fdf3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86c2da64ed7a7f6af14d9fdd54dbb1ace65f683dc22f739afbffa9f1cd6d52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.866469 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.887427 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:38 crc kubenswrapper[4814]: I0227 16:25:38.903065 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:38Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:39 crc kubenswrapper[4814]: I0227 16:25:39.487583 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:39 crc kubenswrapper[4814]: E0227 16:25:39.488357 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:39 crc kubenswrapper[4814]: I0227 16:25:39.487772 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:39 crc kubenswrapper[4814]: E0227 16:25:39.488486 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:39 crc kubenswrapper[4814]: I0227 16:25:39.487883 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:39 crc kubenswrapper[4814]: E0227 16:25:39.488571 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:39 crc kubenswrapper[4814]: I0227 16:25:39.487723 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:39 crc kubenswrapper[4814]: E0227 16:25:39.488662 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.268352 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4m267_56739684-228b-4bf3-bb4e-11f7aaca1890/kube-multus/0.log" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.268414 4814 generic.go:334] "Generic (PLEG): container finished" podID="56739684-228b-4bf3-bb4e-11f7aaca1890" containerID="46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5" exitCode=1 Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.268461 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4m267" event={"ID":"56739684-228b-4bf3-bb4e-11f7aaca1890","Type":"ContainerDied","Data":"46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5"} Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.269057 4814 scope.go:117] "RemoveContainer" containerID="46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.288327 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.311491 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.332298 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.352625 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.378095 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.400058 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36413f7d-bc90-4ab1-8cbd-e9a7a36f92cf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe23574a71dd13340a8be7b9dae77410b55fd1daa11a58eb8ed914fb44cea829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34064bf25c31b4fb36554561a8cf9109191e950be41e22c71e9d4773e0e66ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc29a3f6d746e0eda7be7e511df92aa978993ac600a788820e292603b328434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.416159 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.432362 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8784be35b3b5c5fb9061a38828d5af89d6ee8b91181760938a1fa9c13f3b685b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e61727677ea1f23be24d9615f544f3e4e32330117a45dcd459de41369662b388\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.465603 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.485755 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.486722 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.486726 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:41 crc kubenswrapper[4814]: E0227 16:25:41.486911 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.486749 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:41 crc kubenswrapper[4814]: E0227 16:25:41.487000 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:41 crc kubenswrapper[4814]: E0227 16:25:41.487140 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.487388 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:41 crc kubenswrapper[4814]: E0227 16:25:41.487562 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.501788 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.517056 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45306fad-bea6-4c07-92ba-a910f8e26626\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b7hx9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.533410 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:41Z\\\",\\\"message\\\":\\\"2026-02-27T16:24:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_68d93f49-93e4-4a14-9992-4e4093e4a701\\\\n2026-02-27T16:24:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_68d93f49-93e4-4a14-9992-4e4093e4a701 to /host/opt/cni/bin/\\\\n2026-02-27T16:24:56Z [verbose] multus-daemon started\\\\n2026-02-27T16:24:56Z [verbose] Readiness Indicator file check\\\\n2026-02-27T16:25:41Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.554666 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:23Z\\\",\\\"message\\\":\\\"/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0227 16:25:23.572210 7108 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 16:25:23.572245 7108 factory.go:656] Stopping watch factory\\\\nI0227 16:25:23.572272 7108 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:23.572302 7108 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 16:25:23.572311 7108 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 16:25:23.572406 7108 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.573065 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.591426 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.606171 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.624167 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d657f969-c1a7-4c81-ad14-57a0d0660d46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d1c14a07903094a68c4b62d08ede23259920fb342412bc54a5d1ce0ea92b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14102163d15d2563d01aa8d820f93aa6c36a244adb644431399aa94fea3b568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14102163d15d2563d01aa8d820f93aa6c36a244adb644431399aa94fea3b568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:41 crc kubenswrapper[4814]: I0227 16:25:41.645082 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8269fdbb-d61c-4b25-8fb1-251ee6857851\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22d8635e425d164cd7c63685f0fb8b64a69919a2a687c9ca42321bfed7559ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5ba138e1d2cbeab8ba50d4d02b246c31cb1226615131fa4b074a08d615ca4f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:23:41Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 16:23:10.695873 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 16:23:10.702271 1 observer_polling.go:159] Starting file observer\\\\nI0227 16:23:10.732327 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 16:23:10.742605 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0227 16:23:41.467207 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:40Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501ea405666a49080dd2af178db81a992757c23da06fedd1a6036047880f3f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbf88717d3b766e6b99c0ec623769bd67642fe95222e221d59f15e0e2fdf3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86c2da64ed7a7f6af14d9fdd54dbb1ace65f683dc22f739afbffa9f1cd6d52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:41Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.275180 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4m267_56739684-228b-4bf3-bb4e-11f7aaca1890/kube-multus/0.log" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.275273 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4m267" event={"ID":"56739684-228b-4bf3-bb4e-11f7aaca1890","Type":"ContainerStarted","Data":"c7417cf941b0d6c4cba975b6596d39476bda8e19c0399f28a0288e09bc436549"} Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.298765 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.317443 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8784be35b3b5c5fb9061a38828d5af89d6ee8b91181760938a1fa9c13f3b685b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e61727677ea1f23be24d9615f544f3e4e32330117a45dcd459de41369662b388\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.353767 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.392774 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:23Z\\\",\\\"message\\\":\\\"/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0227 16:25:23.572210 7108 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 16:25:23.572245 7108 factory.go:656] Stopping watch factory\\\\nI0227 16:25:23.572272 7108 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:23.572302 7108 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 16:25:23.572311 7108 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 16:25:23.572406 7108 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.413544 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.432016 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45306fad-bea6-4c07-92ba-a910f8e26626\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b7hx9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.458138 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7417cf941b0d6c4cba975b6596d39476bda8e19c0399f28a0288e09bc436549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:41Z\\\",\\\"message\\\":\\\"2026-02-27T16:24:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_68d93f49-93e4-4a14-9992-4e4093e4a701\\\\n2026-02-27T16:24:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_68d93f49-93e4-4a14-9992-4e4093e4a701 to /host/opt/cni/bin/\\\\n2026-02-27T16:24:56Z [verbose] multus-daemon started\\\\n2026-02-27T16:24:56Z [verbose] Readiness Indicator file check\\\\n2026-02-27T16:25:41Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.484882 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8269fdbb-d61c-4b25-8fb1-251ee6857851\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22d8635e425d164cd7c63685f0fb8b64a69919a2a687c9ca42321bfed7559ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5ba138e1d2cbeab8ba50d4d02b246c31cb1226615131fa4b074a08d615ca4f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:23:41Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 16:23:10.695873 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 16:23:10.702271 1 observer_polling.go:159] Starting file observer\\\\nI0227 16:23:10.732327 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 16:23:10.742605 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0227 16:23:41.467207 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:40Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501ea405666a49080dd2af178db81a992757c23da06fedd1a6036047880f3f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbf88717d3b766e6b99c0ec623769bd67642fe95222e221d59f15e0e2fdf3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86c2da64ed7a7f6af14d9fdd54dbb1ace65f683dc22f739afbffa9f1cd6d52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.501995 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.519143 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.534660 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.549645 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d657f969-c1a7-4c81-ad14-57a0d0660d46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d1c14a07903094a68c4b62d08ede23259920fb342412bc54a5d1ce0ea92b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14102163d15d2563d01aa8d820f93aa6c36a244adb644431399aa94fea3b568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14102163d15d2563d01aa8d820f93aa6c36a244adb644431399aa94fea3b568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.567622 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.584979 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.601006 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.619423 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.635334 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.661406 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:42 crc kubenswrapper[4814]: I0227 16:25:42.681815 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36413f7d-bc90-4ab1-8cbd-e9a7a36f92cf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe23574a71dd13340a8be7b9dae77410b55fd1daa11a58eb8ed914fb44cea829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34064bf25c31b4fb36554561a8cf9109191e950be41e22c71e9d4773e0e66ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc29a3f6d746e0eda7be7e511df92aa978993ac600a788820e292603b328434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:42Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:43 crc kubenswrapper[4814]: I0227 16:25:43.487219 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:43 crc kubenswrapper[4814]: I0227 16:25:43.487431 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:43 crc kubenswrapper[4814]: I0227 16:25:43.487457 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:43 crc kubenswrapper[4814]: E0227 16:25:43.487616 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:43 crc kubenswrapper[4814]: E0227 16:25:43.487769 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:43 crc kubenswrapper[4814]: E0227 16:25:43.487941 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:43 crc kubenswrapper[4814]: I0227 16:25:43.488137 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:43 crc kubenswrapper[4814]: E0227 16:25:43.488298 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:43 crc kubenswrapper[4814]: E0227 16:25:43.611769 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:25:45 crc kubenswrapper[4814]: I0227 16:25:45.487088 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:45 crc kubenswrapper[4814]: I0227 16:25:45.487186 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:45 crc kubenswrapper[4814]: I0227 16:25:45.487233 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:45 crc kubenswrapper[4814]: I0227 16:25:45.487376 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:45 crc kubenswrapper[4814]: E0227 16:25:45.487567 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:45 crc kubenswrapper[4814]: E0227 16:25:45.487709 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:45 crc kubenswrapper[4814]: E0227 16:25:45.487902 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:45 crc kubenswrapper[4814]: E0227 16:25:45.487983 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.486585 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.486720 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:47 crc kubenswrapper[4814]: E0227 16:25:47.486792 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.486829 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.487012 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:47 crc kubenswrapper[4814]: E0227 16:25:47.487052 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:47 crc kubenswrapper[4814]: E0227 16:25:47.487114 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:47 crc kubenswrapper[4814]: E0227 16:25:47.487187 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.657883 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.657964 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.657987 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.658017 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.658038 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:47Z","lastTransitionTime":"2026-02-27T16:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:47 crc kubenswrapper[4814]: E0227 16:25:47.679848 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:47Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.686185 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.686306 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.686328 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.686357 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.686377 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:47Z","lastTransitionTime":"2026-02-27T16:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:47 crc kubenswrapper[4814]: E0227 16:25:47.709608 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:47Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.714242 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.714324 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.714344 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.714370 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.714390 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:47Z","lastTransitionTime":"2026-02-27T16:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:47 crc kubenswrapper[4814]: E0227 16:25:47.735636 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:47Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.741042 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.741101 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.741121 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.741147 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.741164 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:47Z","lastTransitionTime":"2026-02-27T16:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:47 crc kubenswrapper[4814]: E0227 16:25:47.761730 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:47Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.768463 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.768530 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.768551 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.768579 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:47 crc kubenswrapper[4814]: I0227 16:25:47.768598 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:47Z","lastTransitionTime":"2026-02-27T16:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:47 crc kubenswrapper[4814]: E0227 16:25:47.788653 4814 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9eed0068-32e1-4174-93ec-451c02789f32\\\",\\\"systemUUID\\\":\\\"1f184c22-f8d8-46af-88b2-5c278287a2f3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:47Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:47 crc kubenswrapper[4814]: E0227 16:25:47.788878 4814 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.526922 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.549120 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.569428 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8784be35b3b5c5fb9061a38828d5af89d6ee8b91181760938a1fa9c13f3b685b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e61727677ea1f23be24d9615f544f3e4e32330117a45dcd459de41369662b388\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.592067 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7417cf941b0d6c4cba975b6596d39476bda8e19c0399f28a0288e09bc436549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:41Z\\\",\\\"message\\\":\\\"2026-02-27T16:24:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_68d93f49-93e4-4a14-9992-4e4093e4a701\\\\n2026-02-27T16:24:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_68d93f49-93e4-4a14-9992-4e4093e4a701 to /host/opt/cni/bin/\\\\n2026-02-27T16:24:56Z [verbose] multus-daemon started\\\\n2026-02-27T16:24:56Z [verbose] Readiness Indicator file check\\\\n2026-02-27T16:25:41Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: E0227 16:25:48.614207 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.615830 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4619e86c-7eae-4849-89b0-abb44778d871\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:23Z\\\",\\\"message\\\":\\\"/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0227 16:25:23.572210 7108 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 16:25:23.572245 7108 factory.go:656] Stopping watch factory\\\\nI0227 16:25:23.572272 7108 ovnkube.go:599] Stopped ovnkube\\\\nI0227 16:25:23.572302 7108 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 16:25:23.572311 7108 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 16:25:23.572406 7108 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jntcv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zrr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.636606 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5knkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c3aee15-1546-403a-866a-80455c2abb05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc44b25e1413bde149fecaa0e0b0e3e42037e54e5f01726587ae728d7166686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwhk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5knkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.649787 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45306fad-bea6-4c07-92ba-a910f8e26626\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccs8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b7hx9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.666876 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d657f969-c1a7-4c81-ad14-57a0d0660d46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d1c14a07903094a68c4b62d08ede23259920fb342412bc54a5d1ce0ea92b25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f14102163d15d2563d01aa8d820f93aa6c36a244adb644431399aa94fea3b568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14102163d15d2563d01aa8d820f93aa6c36a244adb644431399aa94fea3b568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.689355 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8269fdbb-d61c-4b25-8fb1-251ee6857851\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22d8635e425d164cd7c63685f0fb8b64a69919a2a687c9ca42321bfed7559ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5ba138e1d2cbeab8ba50d4d02b246c31cb1226615131fa4b074a08d615ca4f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:23:41Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 16:23:10.695873 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 16:23:10.702271 1 observer_polling.go:159] Starting file observer\\\\nI0227 16:23:10.732327 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 16:23:10.742605 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0227 16:23:41.467207 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:23:40Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501ea405666a49080dd2af178db81a992757c23da06fedd1a6036047880f3f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbf88717d3b766e6b99c0ec623769bd67642fe95222e221d59f15e0e2fdf3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86c2da64ed7a7f6af14d9fdd54dbb1ace65f683dc22f739afbffa9f1cd6d52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.711694 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba49ba095563d40878a9f87aef378a6b26dfc5741f5d062114995cecbe599fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.730821 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89fbbe036a1eec711a953db1f90788f6ef0a28f008c8c497bbdd3e7e527959b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.752381 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hmxvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e61af0a-b3e1-4131-a0d4-2c550b00cb16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8d2b5b6c9e40e14f32c6d53e172b45463c4f9d59b7680a75012013c0a1154b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9zkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hmxvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.775394 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36413f7d-bc90-4ab1-8cbd-e9a7a36f92cf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe23574a71dd13340a8be7b9dae77410b55fd1daa11a58eb8ed914fb44cea829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34064bf25c31b4fb36554561a8cf9109191e950be41e22c71e9d4773e0e66ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc29a3f6d746e0eda7be7e511df92aa978993ac600a788820e292603b328434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.797161 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.820778 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.842493 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.862342 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.883271 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:48 crc kubenswrapper[4814]: I0227 16:25:48.910764 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:48Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:49 crc kubenswrapper[4814]: I0227 16:25:49.486973 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:49 crc kubenswrapper[4814]: I0227 16:25:49.487168 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:49 crc kubenswrapper[4814]: I0227 16:25:49.487196 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:49 crc kubenswrapper[4814]: I0227 16:25:49.487396 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:49 crc kubenswrapper[4814]: E0227 16:25:49.487682 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:49 crc kubenswrapper[4814]: E0227 16:25:49.487819 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:49 crc kubenswrapper[4814]: E0227 16:25:49.487903 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:49 crc kubenswrapper[4814]: E0227 16:25:49.488018 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:51 crc kubenswrapper[4814]: I0227 16:25:51.487604 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:51 crc kubenswrapper[4814]: I0227 16:25:51.487657 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:51 crc kubenswrapper[4814]: I0227 16:25:51.487695 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:51 crc kubenswrapper[4814]: E0227 16:25:51.487858 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:51 crc kubenswrapper[4814]: E0227 16:25:51.487984 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:51 crc kubenswrapper[4814]: E0227 16:25:51.488127 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:51 crc kubenswrapper[4814]: I0227 16:25:51.488378 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:51 crc kubenswrapper[4814]: E0227 16:25:51.488489 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:53 crc kubenswrapper[4814]: I0227 16:25:53.487505 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:53 crc kubenswrapper[4814]: E0227 16:25:53.488204 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:53 crc kubenswrapper[4814]: I0227 16:25:53.488559 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:53 crc kubenswrapper[4814]: E0227 16:25:53.488663 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:53 crc kubenswrapper[4814]: I0227 16:25:53.490050 4814 scope.go:117] "RemoveContainer" containerID="4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990" Feb 27 16:25:53 crc kubenswrapper[4814]: I0227 16:25:53.490590 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:53 crc kubenswrapper[4814]: E0227 16:25:53.490716 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:53 crc kubenswrapper[4814]: I0227 16:25:53.490914 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:53 crc kubenswrapper[4814]: E0227 16:25:53.491006 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:53 crc kubenswrapper[4814]: E0227 16:25:53.615348 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.327009 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovnkube-controller/2.log" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.331536 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerStarted","Data":"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd"} Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.332527 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.355628 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36413f7d-bc90-4ab1-8cbd-e9a7a36f92cf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe23574a71dd13340a8be7b9dae77410b55fd1daa11a58eb8ed914fb44cea829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34064bf25c31b4fb36554561a8cf9109191e950be41e22c71e9d4773e0e66ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc29a3f6d746e0eda7be7e511df92aa978993ac600a788820e292603b328434\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c208354d3494c007fadd6f19df545e88ac463a951ae50e76ed366f3578b3d81b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.380154 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.403596 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2afe31a4443ee852aff8f6d2bc3171289767735ca0d338b4acdbeac75a1cf199\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b68b70d30142c274ba397fd4bac959d393f4d2ff580c7e2bc8b3e1ee938a3e8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.428494 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.448961 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:33Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.469956 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be713cf4-11a1-42e7-93c6-ab3f748570a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a811ea48111e08a225c30878b88b275404455f2ee80c7fb9baeb9689cb89c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6thr9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nlbk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.496532 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bf0649e-4683-4705-9006-df485b9b5250\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f792ce949ec9774e97282ec90647d5f752a05981edbf57dde9f2c1d6e8765d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de9d459669a6195c6e44d62c8290c89b21eade5078451eaf26fe6aaaa7447304\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8557b8dbb8ff1ca9ce86e4249d791bb66089a75c8420c324e3907dcbd5b7f33b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0dab6defbf67ec8db51fd67bbe94cd737720c3534c0f1a882fdab7ce002b3bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a99bc981ceb93e03cbc7007d2014843828ed01d50f53caf7349ee4cb5c16e8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1e3dc4fe44b90f9bf4c228753c6f1beccb3dba3b7eac9e7a21baad5fcacc28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43cc774d2068001b83fab8893ab713d972b335e5ddcabda90f922025b400bc5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ms7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sh8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.532467 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49c4f6d-f042-43aa-82d7-028781e9f87d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96d1411739f9e541f4da3caba66077f020d77bc5a7b836489cd9f1a26884dcf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f4a0104b92fd8e06c4251a7d1c30ced9f95e2f3fb82bbd7f69550493837c1a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcb7848f3e3c1d2bd87f02f4bc2729f7c58a3b6e97d22c6d3bba561cb3a4d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80c3f31b660d31c07d7993a364963757d1d8ec35a6d9c443030de2524a18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edee9f2d52c79351f06be48d4c7f8010ac5dac04be5aa96b6956ba4f085f5733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://279d36cd6918b9141bbc4d42c942d70db4969f389bab702853d5b82f5890a8d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbd1f8e9b95a9e1f2d79ee8aacd2b53a319a012bb2e5bb79536fae0d57770915\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95d921297a825ff951b6f5a066ac4c2be7e83a4121d5dc032d1bdcb65e06224\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.568682 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"929bfc6f-c721-4f45-9674-fc3f151c3f6e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:23:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T16:24:19Z\\\",\\\"message\\\":\\\"file observer\\\\nW0227 16:24:19.118522 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0227 16:24:19.118719 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0227 16:24:19.119531 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-562751553/tls.crt::/tmp/serving-cert-562751553/tls.key\\\\\\\"\\\\nI0227 16:24:19.382692 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0227 16:24:19.388181 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0227 16:24:19.388220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0227 16:24:19.388287 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0227 16:24:19.388304 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0227 16:24:19.400924 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0227 16:24:19.400975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400983 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0227 16:24:19.400989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0227 16:24:19.400994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 16:24:19.400998 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 16:24:19.401002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 16:24:19.401375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0227 16:24:19.407180 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:23:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T16:23:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T16:23:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:23:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.585753 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"444e44da-4a93-4115-9be8-7a5a83db72a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8784be35b3b5c5fb9061a38828d5af89d6ee8b91181760938a1fa9c13f3b685b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e61727677ea1f23be24d9615f544f3e4e32330117a45dcd459de41369662b388\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trffm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:25:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h2s86\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.607237 4814 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4m267" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56739684-228b-4bf3-bb4e-11f7aaca1890\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:24:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T16:25:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7417cf941b0d6c4cba975b6596d39476bda8e19c0399f28a0288e09bc436549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T16:25:41Z\\\",\\\"message\\\":\\\"2026-02-27T16:24:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_68d93f49-93e4-4a14-9992-4e4093e4a701\\\\n2026-02-27T16:24:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_68d93f49-93e4-4a14-9992-4e4093e4a701 to /host/opt/cni/bin/\\\\n2026-02-27T16:24:56Z [verbose] multus-daemon started\\\\n2026-02-27T16:24:56Z [verbose] Readiness Indicator file check\\\\n2026-02-27T16:25:41Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T16:24:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T16:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kdjfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T16:24:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4m267\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T16:25:54Z is after 2025-08-24T17:21:41Z" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.655526 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podStartSLOduration=101.655501276 podStartE2EDuration="1m41.655501276s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:25:54.654714137 +0000 UTC m=+167.107339017" watchObservedRunningTime="2026-02-27 16:25:54.655501276 +0000 UTC m=+167.108126146" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.681040 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-5knkr" podStartSLOduration=101.681002506 podStartE2EDuration="1m41.681002506s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:25:54.668599253 +0000 UTC m=+167.121224103" watchObservedRunningTime="2026-02-27 16:25:54.681002506 +0000 UTC m=+167.133627376" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.698029 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=17.698002719 podStartE2EDuration="17.698002719s" podCreationTimestamp="2026-02-27 16:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:25:54.697655606 +0000 UTC m=+167.150280536" watchObservedRunningTime="2026-02-27 16:25:54.698002719 +0000 UTC m=+167.150627559" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.741924 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=36.741903563 podStartE2EDuration="36.741903563s" podCreationTimestamp="2026-02-27 16:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:25:54.717977482 +0000 UTC m=+167.170602322" watchObservedRunningTime="2026-02-27 16:25:54.741903563 +0000 UTC m=+167.194528403" Feb 27 16:25:54 crc kubenswrapper[4814]: I0227 16:25:54.784475 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-hmxvd" podStartSLOduration=101.784453418 podStartE2EDuration="1m41.784453418s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:25:54.784353124 +0000 UTC m=+167.236977984" watchObservedRunningTime="2026-02-27 16:25:54.784453418 +0000 UTC m=+167.237078258" Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.338578 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovnkube-controller/3.log" Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.339795 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovnkube-controller/2.log" Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.344552 4814 generic.go:334] "Generic (PLEG): container finished" podID="4619e86c-7eae-4849-89b0-abb44778d871" containerID="17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd" exitCode=1 Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.344649 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerDied","Data":"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd"} Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.344757 4814 scope.go:117] "RemoveContainer" containerID="4f29f762b7d416e674ffcf53f9d0387bd793a080185e7ee2c869903ce1127990" Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.345972 4814 scope.go:117] "RemoveContainer" containerID="17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd" Feb 27 16:25:55 crc kubenswrapper[4814]: E0227 16:25:55.346493 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.370563 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=39.370534099 podStartE2EDuration="39.370534099s" podCreationTimestamp="2026-02-27 16:25:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:25:55.369873886 +0000 UTC m=+167.822498756" watchObservedRunningTime="2026-02-27 16:25:55.370534099 +0000 UTC m=+167.823158949" Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.481123 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podStartSLOduration=102.481098167 podStartE2EDuration="1m42.481098167s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:25:55.480578727 +0000 UTC m=+167.933203617" watchObservedRunningTime="2026-02-27 16:25:55.481098167 +0000 UTC m=+167.933723007" Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.486626 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.486898 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:55 crc kubenswrapper[4814]: E0227 16:25:55.486949 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.486684 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:55 crc kubenswrapper[4814]: E0227 16:25:55.487182 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:55 crc kubenswrapper[4814]: E0227 16:25:55.487510 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.487538 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:55 crc kubenswrapper[4814]: E0227 16:25:55.487685 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.510629 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-sh8gf" podStartSLOduration=102.510599245 podStartE2EDuration="1m42.510599245s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:25:55.508796768 +0000 UTC m=+167.961421638" watchObservedRunningTime="2026-02-27 16:25:55.510599245 +0000 UTC m=+167.963224115" Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.568769 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=77.56873964 podStartE2EDuration="1m17.56873964s" podCreationTimestamp="2026-02-27 16:24:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:25:55.567854427 +0000 UTC m=+168.020479297" watchObservedRunningTime="2026-02-27 16:25:55.56873964 +0000 UTC m=+168.021364490" Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.568981 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=66.568972538 podStartE2EDuration="1m6.568972538s" podCreationTimestamp="2026-02-27 16:24:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:25:55.54942932 +0000 UTC m=+168.002054150" watchObservedRunningTime="2026-02-27 16:25:55.568972538 +0000 UTC m=+168.021597388" Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.582280 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h2s86" podStartSLOduration=102.582230072 podStartE2EDuration="1m42.582230072s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:25:55.581927921 +0000 UTC m=+168.034552791" watchObservedRunningTime="2026-02-27 16:25:55.582230072 +0000 UTC m=+168.034854912" Feb 27 16:25:55 crc kubenswrapper[4814]: I0227 16:25:55.597872 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-4m267" podStartSLOduration=102.597846993 podStartE2EDuration="1m42.597846993s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:25:55.597846203 +0000 UTC m=+168.050471043" watchObservedRunningTime="2026-02-27 16:25:55.597846993 +0000 UTC m=+168.050471833" Feb 27 16:25:56 crc kubenswrapper[4814]: I0227 16:25:56.352019 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovnkube-controller/3.log" Feb 27 16:25:56 crc kubenswrapper[4814]: I0227 16:25:56.358406 4814 scope.go:117] "RemoveContainer" containerID="17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd" Feb 27 16:25:56 crc kubenswrapper[4814]: E0227 16:25:56.358712 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" Feb 27 16:25:57 crc kubenswrapper[4814]: I0227 16:25:57.486757 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:57 crc kubenswrapper[4814]: I0227 16:25:57.486976 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:57 crc kubenswrapper[4814]: I0227 16:25:57.487010 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:57 crc kubenswrapper[4814]: I0227 16:25:57.487080 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:57 crc kubenswrapper[4814]: E0227 16:25:57.487544 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:57 crc kubenswrapper[4814]: E0227 16:25:57.487629 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:25:57 crc kubenswrapper[4814]: E0227 16:25:57.487865 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:57 crc kubenswrapper[4814]: E0227 16:25:57.488023 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:57 crc kubenswrapper[4814]: I0227 16:25:57.864594 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 16:25:57 crc kubenswrapper[4814]: I0227 16:25:57.864685 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 16:25:57 crc kubenswrapper[4814]: I0227 16:25:57.864701 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 16:25:57 crc kubenswrapper[4814]: I0227 16:25:57.864719 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 16:25:57 crc kubenswrapper[4814]: I0227 16:25:57.864733 4814 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T16:25:57Z","lastTransitionTime":"2026-02-27T16:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 16:25:57 crc kubenswrapper[4814]: I0227 16:25:57.936076 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs"] Feb 27 16:25:57 crc kubenswrapper[4814]: I0227 16:25:57.936490 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:57 crc kubenswrapper[4814]: I0227 16:25:57.939739 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 27 16:25:57 crc kubenswrapper[4814]: I0227 16:25:57.940011 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 27 16:25:57 crc kubenswrapper[4814]: I0227 16:25:57.940192 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 27 16:25:57 crc kubenswrapper[4814]: I0227 16:25:57.942984 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.070348 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/201a12b2-e1b4-494f-b724-12f048c3ca92-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-h2grs\" (UID: \"201a12b2-e1b4-494f-b724-12f048c3ca92\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.070428 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/201a12b2-e1b4-494f-b724-12f048c3ca92-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-h2grs\" (UID: \"201a12b2-e1b4-494f-b724-12f048c3ca92\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.070467 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/201a12b2-e1b4-494f-b724-12f048c3ca92-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-h2grs\" (UID: \"201a12b2-e1b4-494f-b724-12f048c3ca92\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.070500 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/201a12b2-e1b4-494f-b724-12f048c3ca92-service-ca\") pod \"cluster-version-operator-5c965bbfc6-h2grs\" (UID: \"201a12b2-e1b4-494f-b724-12f048c3ca92\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.070598 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/201a12b2-e1b4-494f-b724-12f048c3ca92-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-h2grs\" (UID: \"201a12b2-e1b4-494f-b724-12f048c3ca92\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.171444 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/201a12b2-e1b4-494f-b724-12f048c3ca92-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-h2grs\" (UID: \"201a12b2-e1b4-494f-b724-12f048c3ca92\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.171533 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/201a12b2-e1b4-494f-b724-12f048c3ca92-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-h2grs\" (UID: \"201a12b2-e1b4-494f-b724-12f048c3ca92\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.171572 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/201a12b2-e1b4-494f-b724-12f048c3ca92-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-h2grs\" (UID: \"201a12b2-e1b4-494f-b724-12f048c3ca92\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.171606 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/201a12b2-e1b4-494f-b724-12f048c3ca92-service-ca\") pod \"cluster-version-operator-5c965bbfc6-h2grs\" (UID: \"201a12b2-e1b4-494f-b724-12f048c3ca92\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.171679 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/201a12b2-e1b4-494f-b724-12f048c3ca92-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-h2grs\" (UID: \"201a12b2-e1b4-494f-b724-12f048c3ca92\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.171681 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/201a12b2-e1b4-494f-b724-12f048c3ca92-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-h2grs\" (UID: \"201a12b2-e1b4-494f-b724-12f048c3ca92\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.171809 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/201a12b2-e1b4-494f-b724-12f048c3ca92-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-h2grs\" (UID: \"201a12b2-e1b4-494f-b724-12f048c3ca92\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.173586 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/201a12b2-e1b4-494f-b724-12f048c3ca92-service-ca\") pod \"cluster-version-operator-5c965bbfc6-h2grs\" (UID: \"201a12b2-e1b4-494f-b724-12f048c3ca92\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.180321 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/201a12b2-e1b4-494f-b724-12f048c3ca92-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-h2grs\" (UID: \"201a12b2-e1b4-494f-b724-12f048c3ca92\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.203472 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/201a12b2-e1b4-494f-b724-12f048c3ca92-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-h2grs\" (UID: \"201a12b2-e1b4-494f-b724-12f048c3ca92\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.258579 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" Feb 27 16:25:58 crc kubenswrapper[4814]: W0227 16:25:58.282696 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod201a12b2_e1b4_494f_b724_12f048c3ca92.slice/crio-4aab09855769656e0cf272b60e01ed42aaab172d189b2fe7f34b7c1c3e29c4ae WatchSource:0}: Error finding container 4aab09855769656e0cf272b60e01ed42aaab172d189b2fe7f34b7c1c3e29c4ae: Status 404 returned error can't find the container with id 4aab09855769656e0cf272b60e01ed42aaab172d189b2fe7f34b7c1c3e29c4ae Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.368583 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" event={"ID":"201a12b2-e1b4-494f-b724-12f048c3ca92","Type":"ContainerStarted","Data":"4aab09855769656e0cf272b60e01ed42aaab172d189b2fe7f34b7c1c3e29c4ae"} Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.526833 4814 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 27 16:25:58 crc kubenswrapper[4814]: I0227 16:25:58.540550 4814 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 27 16:25:58 crc kubenswrapper[4814]: E0227 16:25:58.616874 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:25:59 crc kubenswrapper[4814]: I0227 16:25:59.375077 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" event={"ID":"201a12b2-e1b4-494f-b724-12f048c3ca92","Type":"ContainerStarted","Data":"dd66c1e38171a1c35536446d1850f59f3cb1e6ab35530969820f6394893a7980"} Feb 27 16:25:59 crc kubenswrapper[4814]: I0227 16:25:59.396745 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h2grs" podStartSLOduration=106.396725731 podStartE2EDuration="1m46.396725731s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:25:59.39642112 +0000 UTC m=+171.849045990" watchObservedRunningTime="2026-02-27 16:25:59.396725731 +0000 UTC m=+171.849350601" Feb 27 16:25:59 crc kubenswrapper[4814]: I0227 16:25:59.487096 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:25:59 crc kubenswrapper[4814]: I0227 16:25:59.487186 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:25:59 crc kubenswrapper[4814]: I0227 16:25:59.487112 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:25:59 crc kubenswrapper[4814]: E0227 16:25:59.487331 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:25:59 crc kubenswrapper[4814]: E0227 16:25:59.487504 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:25:59 crc kubenswrapper[4814]: E0227 16:25:59.487638 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:25:59 crc kubenswrapper[4814]: I0227 16:25:59.488082 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:25:59 crc kubenswrapper[4814]: E0227 16:25:59.488498 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:01 crc kubenswrapper[4814]: I0227 16:26:01.487218 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:01 crc kubenswrapper[4814]: I0227 16:26:01.487286 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:01 crc kubenswrapper[4814]: I0227 16:26:01.487359 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:01 crc kubenswrapper[4814]: E0227 16:26:01.487898 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:01 crc kubenswrapper[4814]: E0227 16:26:01.488235 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:01 crc kubenswrapper[4814]: E0227 16:26:01.488401 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:01 crc kubenswrapper[4814]: I0227 16:26:01.488742 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:01 crc kubenswrapper[4814]: E0227 16:26:01.488869 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:03 crc kubenswrapper[4814]: I0227 16:26:03.487555 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:03 crc kubenswrapper[4814]: I0227 16:26:03.487683 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:03 crc kubenswrapper[4814]: I0227 16:26:03.487686 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:03 crc kubenswrapper[4814]: I0227 16:26:03.487815 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:03 crc kubenswrapper[4814]: E0227 16:26:03.487833 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:03 crc kubenswrapper[4814]: E0227 16:26:03.487978 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:03 crc kubenswrapper[4814]: E0227 16:26:03.488123 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:03 crc kubenswrapper[4814]: E0227 16:26:03.488224 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:03 crc kubenswrapper[4814]: E0227 16:26:03.618819 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:26:05 crc kubenswrapper[4814]: I0227 16:26:05.487645 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:05 crc kubenswrapper[4814]: I0227 16:26:05.487725 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:05 crc kubenswrapper[4814]: I0227 16:26:05.487666 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:05 crc kubenswrapper[4814]: I0227 16:26:05.487698 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:05 crc kubenswrapper[4814]: E0227 16:26:05.487848 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:05 crc kubenswrapper[4814]: E0227 16:26:05.487953 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:05 crc kubenswrapper[4814]: E0227 16:26:05.488136 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:05 crc kubenswrapper[4814]: E0227 16:26:05.488307 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:07 crc kubenswrapper[4814]: I0227 16:26:07.486873 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:07 crc kubenswrapper[4814]: I0227 16:26:07.486951 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:07 crc kubenswrapper[4814]: I0227 16:26:07.486947 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:07 crc kubenswrapper[4814]: I0227 16:26:07.487031 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:07 crc kubenswrapper[4814]: E0227 16:26:07.487243 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:07 crc kubenswrapper[4814]: E0227 16:26:07.487419 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:07 crc kubenswrapper[4814]: E0227 16:26:07.487574 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:07 crc kubenswrapper[4814]: E0227 16:26:07.487695 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:08 crc kubenswrapper[4814]: E0227 16:26:08.621539 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:26:09 crc kubenswrapper[4814]: I0227 16:26:09.486630 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:09 crc kubenswrapper[4814]: I0227 16:26:09.486637 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:09 crc kubenswrapper[4814]: I0227 16:26:09.486673 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:09 crc kubenswrapper[4814]: I0227 16:26:09.487525 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:09 crc kubenswrapper[4814]: E0227 16:26:09.487731 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:09 crc kubenswrapper[4814]: E0227 16:26:09.487647 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:09 crc kubenswrapper[4814]: E0227 16:26:09.488157 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:09 crc kubenswrapper[4814]: E0227 16:26:09.488289 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:10 crc kubenswrapper[4814]: I0227 16:26:10.011787 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs\") pod \"network-metrics-daemon-b7hx9\" (UID: \"45306fad-bea6-4c07-92ba-a910f8e26626\") " pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:10 crc kubenswrapper[4814]: E0227 16:26:10.012115 4814 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 16:26:10 crc kubenswrapper[4814]: E0227 16:26:10.012244 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs podName:45306fad-bea6-4c07-92ba-a910f8e26626 nodeName:}" failed. No retries permitted until 2026-02-27 16:27:14.012207468 +0000 UTC m=+246.464832338 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs") pod "network-metrics-daemon-b7hx9" (UID: "45306fad-bea6-4c07-92ba-a910f8e26626") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 16:26:11 crc kubenswrapper[4814]: I0227 16:26:11.487334 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:11 crc kubenswrapper[4814]: I0227 16:26:11.487403 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:11 crc kubenswrapper[4814]: I0227 16:26:11.487359 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:11 crc kubenswrapper[4814]: E0227 16:26:11.487507 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:11 crc kubenswrapper[4814]: I0227 16:26:11.488571 4814 scope.go:117] "RemoveContainer" containerID="17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd" Feb 27 16:26:11 crc kubenswrapper[4814]: E0227 16:26:11.488811 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" Feb 27 16:26:11 crc kubenswrapper[4814]: E0227 16:26:11.488949 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:11 crc kubenswrapper[4814]: I0227 16:26:11.488833 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:11 crc kubenswrapper[4814]: E0227 16:26:11.489150 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:11 crc kubenswrapper[4814]: E0227 16:26:11.489212 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:13 crc kubenswrapper[4814]: I0227 16:26:13.487680 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:13 crc kubenswrapper[4814]: I0227 16:26:13.487763 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:13 crc kubenswrapper[4814]: I0227 16:26:13.487802 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:13 crc kubenswrapper[4814]: E0227 16:26:13.487913 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:13 crc kubenswrapper[4814]: I0227 16:26:13.487935 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:13 crc kubenswrapper[4814]: E0227 16:26:13.488094 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:13 crc kubenswrapper[4814]: E0227 16:26:13.488294 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:13 crc kubenswrapper[4814]: E0227 16:26:13.488540 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:13 crc kubenswrapper[4814]: E0227 16:26:13.622910 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:26:15 crc kubenswrapper[4814]: I0227 16:26:15.487291 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:15 crc kubenswrapper[4814]: I0227 16:26:15.487331 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:15 crc kubenswrapper[4814]: I0227 16:26:15.487361 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:15 crc kubenswrapper[4814]: I0227 16:26:15.487302 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:15 crc kubenswrapper[4814]: E0227 16:26:15.487541 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:15 crc kubenswrapper[4814]: E0227 16:26:15.487640 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:15 crc kubenswrapper[4814]: E0227 16:26:15.487766 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:15 crc kubenswrapper[4814]: E0227 16:26:15.487923 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:17 crc kubenswrapper[4814]: I0227 16:26:17.486971 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:17 crc kubenswrapper[4814]: I0227 16:26:17.487053 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:17 crc kubenswrapper[4814]: E0227 16:26:17.487220 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:17 crc kubenswrapper[4814]: E0227 16:26:17.487409 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:17 crc kubenswrapper[4814]: I0227 16:26:17.488422 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:17 crc kubenswrapper[4814]: E0227 16:26:17.488537 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:17 crc kubenswrapper[4814]: I0227 16:26:17.488619 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:17 crc kubenswrapper[4814]: E0227 16:26:17.488833 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:18 crc kubenswrapper[4814]: E0227 16:26:18.624589 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:26:19 crc kubenswrapper[4814]: I0227 16:26:19.487030 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:19 crc kubenswrapper[4814]: I0227 16:26:19.487120 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:19 crc kubenswrapper[4814]: E0227 16:26:19.487181 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:19 crc kubenswrapper[4814]: I0227 16:26:19.487286 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:19 crc kubenswrapper[4814]: E0227 16:26:19.487365 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:19 crc kubenswrapper[4814]: I0227 16:26:19.487294 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:19 crc kubenswrapper[4814]: E0227 16:26:19.487517 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:19 crc kubenswrapper[4814]: E0227 16:26:19.487544 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:21 crc kubenswrapper[4814]: I0227 16:26:21.486767 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:21 crc kubenswrapper[4814]: I0227 16:26:21.486805 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:21 crc kubenswrapper[4814]: E0227 16:26:21.487041 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:21 crc kubenswrapper[4814]: I0227 16:26:21.487154 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:21 crc kubenswrapper[4814]: E0227 16:26:21.487480 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:21 crc kubenswrapper[4814]: E0227 16:26:21.487588 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:21 crc kubenswrapper[4814]: I0227 16:26:21.488609 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:21 crc kubenswrapper[4814]: E0227 16:26:21.488848 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:23 crc kubenswrapper[4814]: I0227 16:26:23.487117 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:23 crc kubenswrapper[4814]: I0227 16:26:23.487203 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:23 crc kubenswrapper[4814]: E0227 16:26:23.487279 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:23 crc kubenswrapper[4814]: I0227 16:26:23.487331 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:23 crc kubenswrapper[4814]: I0227 16:26:23.487379 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:23 crc kubenswrapper[4814]: E0227 16:26:23.487511 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:23 crc kubenswrapper[4814]: E0227 16:26:23.487644 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:23 crc kubenswrapper[4814]: E0227 16:26:23.487789 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:23 crc kubenswrapper[4814]: E0227 16:26:23.625876 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:26:25 crc kubenswrapper[4814]: I0227 16:26:25.487124 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:25 crc kubenswrapper[4814]: I0227 16:26:25.487284 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:25 crc kubenswrapper[4814]: I0227 16:26:25.487377 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:25 crc kubenswrapper[4814]: E0227 16:26:25.487860 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:25 crc kubenswrapper[4814]: I0227 16:26:25.487598 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:25 crc kubenswrapper[4814]: E0227 16:26:25.488525 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:25 crc kubenswrapper[4814]: E0227 16:26:25.489341 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:25 crc kubenswrapper[4814]: E0227 16:26:25.489484 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:25 crc kubenswrapper[4814]: I0227 16:26:25.489941 4814 scope.go:117] "RemoveContainer" containerID="17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd" Feb 27 16:26:25 crc kubenswrapper[4814]: E0227 16:26:25.490309 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7zrr2_openshift-ovn-kubernetes(4619e86c-7eae-4849-89b0-abb44778d871)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" Feb 27 16:26:27 crc kubenswrapper[4814]: I0227 16:26:27.486764 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:27 crc kubenswrapper[4814]: I0227 16:26:27.486805 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:27 crc kubenswrapper[4814]: I0227 16:26:27.486764 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:27 crc kubenswrapper[4814]: I0227 16:26:27.486915 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:27 crc kubenswrapper[4814]: E0227 16:26:27.486935 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:27 crc kubenswrapper[4814]: E0227 16:26:27.487114 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:27 crc kubenswrapper[4814]: E0227 16:26:27.487323 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:27 crc kubenswrapper[4814]: E0227 16:26:27.487506 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:27 crc kubenswrapper[4814]: I0227 16:26:27.491940 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4m267_56739684-228b-4bf3-bb4e-11f7aaca1890/kube-multus/1.log" Feb 27 16:26:27 crc kubenswrapper[4814]: I0227 16:26:27.492615 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4m267_56739684-228b-4bf3-bb4e-11f7aaca1890/kube-multus/0.log" Feb 27 16:26:27 crc kubenswrapper[4814]: I0227 16:26:27.492681 4814 generic.go:334] "Generic (PLEG): container finished" podID="56739684-228b-4bf3-bb4e-11f7aaca1890" containerID="c7417cf941b0d6c4cba975b6596d39476bda8e19c0399f28a0288e09bc436549" exitCode=1 Feb 27 16:26:27 crc kubenswrapper[4814]: I0227 16:26:27.492716 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4m267" event={"ID":"56739684-228b-4bf3-bb4e-11f7aaca1890","Type":"ContainerDied","Data":"c7417cf941b0d6c4cba975b6596d39476bda8e19c0399f28a0288e09bc436549"} Feb 27 16:26:27 crc kubenswrapper[4814]: I0227 16:26:27.492755 4814 scope.go:117] "RemoveContainer" containerID="46916a6545b28ae268ffcaf43c11aa864960eb11aaf7313aee496661a914aed5" Feb 27 16:26:27 crc kubenswrapper[4814]: I0227 16:26:27.493429 4814 scope.go:117] "RemoveContainer" containerID="c7417cf941b0d6c4cba975b6596d39476bda8e19c0399f28a0288e09bc436549" Feb 27 16:26:27 crc kubenswrapper[4814]: E0227 16:26:27.493696 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-4m267_openshift-multus(56739684-228b-4bf3-bb4e-11f7aaca1890)\"" pod="openshift-multus/multus-4m267" podUID="56739684-228b-4bf3-bb4e-11f7aaca1890" Feb 27 16:26:28 crc kubenswrapper[4814]: I0227 16:26:28.498023 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4m267_56739684-228b-4bf3-bb4e-11f7aaca1890/kube-multus/1.log" Feb 27 16:26:28 crc kubenswrapper[4814]: E0227 16:26:28.626989 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:26:29 crc kubenswrapper[4814]: I0227 16:26:29.487605 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:29 crc kubenswrapper[4814]: I0227 16:26:29.487652 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:29 crc kubenswrapper[4814]: I0227 16:26:29.487718 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:29 crc kubenswrapper[4814]: E0227 16:26:29.487790 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:29 crc kubenswrapper[4814]: I0227 16:26:29.487811 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:29 crc kubenswrapper[4814]: E0227 16:26:29.487903 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:29 crc kubenswrapper[4814]: E0227 16:26:29.487993 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:29 crc kubenswrapper[4814]: E0227 16:26:29.488059 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:31 crc kubenswrapper[4814]: I0227 16:26:31.487506 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:31 crc kubenswrapper[4814]: I0227 16:26:31.487543 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:31 crc kubenswrapper[4814]: E0227 16:26:31.487680 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:31 crc kubenswrapper[4814]: I0227 16:26:31.487773 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:31 crc kubenswrapper[4814]: E0227 16:26:31.487920 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:31 crc kubenswrapper[4814]: E0227 16:26:31.488225 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:31 crc kubenswrapper[4814]: I0227 16:26:31.488492 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:31 crc kubenswrapper[4814]: E0227 16:26:31.488627 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:33 crc kubenswrapper[4814]: I0227 16:26:33.486732 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:33 crc kubenswrapper[4814]: I0227 16:26:33.486788 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:33 crc kubenswrapper[4814]: I0227 16:26:33.486768 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:33 crc kubenswrapper[4814]: I0227 16:26:33.486732 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:33 crc kubenswrapper[4814]: E0227 16:26:33.487016 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:33 crc kubenswrapper[4814]: E0227 16:26:33.487160 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:33 crc kubenswrapper[4814]: E0227 16:26:33.487363 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:33 crc kubenswrapper[4814]: E0227 16:26:33.487429 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:33 crc kubenswrapper[4814]: E0227 16:26:33.628381 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:26:35 crc kubenswrapper[4814]: I0227 16:26:35.487464 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:35 crc kubenswrapper[4814]: I0227 16:26:35.487501 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:35 crc kubenswrapper[4814]: I0227 16:26:35.487574 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:35 crc kubenswrapper[4814]: I0227 16:26:35.487665 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:35 crc kubenswrapper[4814]: E0227 16:26:35.487824 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:35 crc kubenswrapper[4814]: E0227 16:26:35.488048 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:35 crc kubenswrapper[4814]: E0227 16:26:35.488196 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:35 crc kubenswrapper[4814]: E0227 16:26:35.488322 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:37 crc kubenswrapper[4814]: I0227 16:26:37.487640 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:37 crc kubenswrapper[4814]: I0227 16:26:37.487737 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:37 crc kubenswrapper[4814]: E0227 16:26:37.487810 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:37 crc kubenswrapper[4814]: I0227 16:26:37.487832 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:37 crc kubenswrapper[4814]: I0227 16:26:37.487873 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:37 crc kubenswrapper[4814]: E0227 16:26:37.488014 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:37 crc kubenswrapper[4814]: E0227 16:26:37.488157 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:37 crc kubenswrapper[4814]: E0227 16:26:37.488382 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:38 crc kubenswrapper[4814]: E0227 16:26:38.629198 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:26:39 crc kubenswrapper[4814]: I0227 16:26:39.487527 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:39 crc kubenswrapper[4814]: I0227 16:26:39.487584 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:39 crc kubenswrapper[4814]: I0227 16:26:39.487629 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:39 crc kubenswrapper[4814]: E0227 16:26:39.487650 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:39 crc kubenswrapper[4814]: I0227 16:26:39.487758 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:39 crc kubenswrapper[4814]: E0227 16:26:39.487872 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:39 crc kubenswrapper[4814]: E0227 16:26:39.488235 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:39 crc kubenswrapper[4814]: E0227 16:26:39.488598 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:39 crc kubenswrapper[4814]: I0227 16:26:39.489108 4814 scope.go:117] "RemoveContainer" containerID="c7417cf941b0d6c4cba975b6596d39476bda8e19c0399f28a0288e09bc436549" Feb 27 16:26:40 crc kubenswrapper[4814]: I0227 16:26:40.488200 4814 scope.go:117] "RemoveContainer" containerID="17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd" Feb 27 16:26:40 crc kubenswrapper[4814]: I0227 16:26:40.556452 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4m267_56739684-228b-4bf3-bb4e-11f7aaca1890/kube-multus/1.log" Feb 27 16:26:40 crc kubenswrapper[4814]: I0227 16:26:40.557009 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4m267" event={"ID":"56739684-228b-4bf3-bb4e-11f7aaca1890","Type":"ContainerStarted","Data":"d7cc356e7ca5938965550870658530bc06e9d97a37f7b465b73ef82649dcd82e"} Feb 27 16:26:41 crc kubenswrapper[4814]: I0227 16:26:41.487544 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:41 crc kubenswrapper[4814]: I0227 16:26:41.487607 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.487724 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:41 crc kubenswrapper[4814]: I0227 16:26:41.487747 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:41 crc kubenswrapper[4814]: I0227 16:26:41.487816 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.487862 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.488018 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.488080 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:41 crc kubenswrapper[4814]: I0227 16:26:41.506037 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:41 crc kubenswrapper[4814]: I0227 16:26:41.506183 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.506268 4814 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.506389 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:28:43.506366554 +0000 UTC m=+335.958991594 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 16:26:41 crc kubenswrapper[4814]: I0227 16:26:41.506284 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.506430 4814 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:26:41 crc kubenswrapper[4814]: I0227 16:26:41.506491 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.506511 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 16:28:43.506487739 +0000 UTC m=+335.959112599 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.506516 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.506566 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.506593 4814 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.506663 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.506679 4814 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.506692 4814 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.506722 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 16:28:43.506685345 +0000 UTC m=+335.959310365 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.506773 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 16:28:43.506750258 +0000 UTC m=+335.959375348 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 16:26:41 crc kubenswrapper[4814]: I0227 16:26:41.563429 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovnkube-controller/3.log" Feb 27 16:26:41 crc kubenswrapper[4814]: I0227 16:26:41.567239 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerStarted","Data":"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a"} Feb 27 16:26:41 crc kubenswrapper[4814]: I0227 16:26:41.567652 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:26:41 crc kubenswrapper[4814]: I0227 16:26:41.633223 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-b7hx9"] Feb 27 16:26:41 crc kubenswrapper[4814]: I0227 16:26:41.633465 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.633678 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:41 crc kubenswrapper[4814]: I0227 16:26:41.710472 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:41 crc kubenswrapper[4814]: E0227 16:26:41.710786 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:28:43.710740942 +0000 UTC m=+336.163365782 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:43 crc kubenswrapper[4814]: I0227 16:26:43.487437 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:43 crc kubenswrapper[4814]: I0227 16:26:43.487555 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:43 crc kubenswrapper[4814]: I0227 16:26:43.487651 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:43 crc kubenswrapper[4814]: E0227 16:26:43.487628 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:43 crc kubenswrapper[4814]: E0227 16:26:43.487793 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:43 crc kubenswrapper[4814]: I0227 16:26:43.487928 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:43 crc kubenswrapper[4814]: E0227 16:26:43.487983 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:43 crc kubenswrapper[4814]: E0227 16:26:43.488191 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:43 crc kubenswrapper[4814]: E0227 16:26:43.631290 4814 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 16:26:45 crc kubenswrapper[4814]: I0227 16:26:45.487563 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:45 crc kubenswrapper[4814]: I0227 16:26:45.487703 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:45 crc kubenswrapper[4814]: I0227 16:26:45.487603 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:45 crc kubenswrapper[4814]: I0227 16:26:45.487563 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:45 crc kubenswrapper[4814]: E0227 16:26:45.487928 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:45 crc kubenswrapper[4814]: E0227 16:26:45.488039 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:45 crc kubenswrapper[4814]: E0227 16:26:45.488277 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:45 crc kubenswrapper[4814]: E0227 16:26:45.488522 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:47 crc kubenswrapper[4814]: I0227 16:26:47.486624 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:47 crc kubenswrapper[4814]: I0227 16:26:47.486676 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:47 crc kubenswrapper[4814]: I0227 16:26:47.486706 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:47 crc kubenswrapper[4814]: E0227 16:26:47.486807 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:26:47 crc kubenswrapper[4814]: I0227 16:26:47.486898 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:47 crc kubenswrapper[4814]: E0227 16:26:47.487092 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b7hx9" podUID="45306fad-bea6-4c07-92ba-a910f8e26626" Feb 27 16:26:47 crc kubenswrapper[4814]: E0227 16:26:47.487213 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 16:26:47 crc kubenswrapper[4814]: E0227 16:26:47.487374 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.039698 4814 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.094866 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ndc8k"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.095424 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.100114 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jds5c"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.100517 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4rvqh"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.100810 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.101199 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.102072 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.103678 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4rvqh" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.104102 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.104193 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.104204 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.104310 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.104440 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.104498 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.104719 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.106570 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-k4p69"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.107136 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.108809 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.109700 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.112177 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.112706 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.114495 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pbsmc"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.115954 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.116748 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.117726 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-7x2qp"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.118436 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.117751 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.119874 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.121325 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9vmj5"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.122134 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.124389 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.124798 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.125333 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.125462 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.125819 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.125909 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.125939 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.126004 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.128282 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.128357 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.128403 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.128551 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.129690 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.134353 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4m8vp"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.135054 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pwdrv"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.135848 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-pwdrv" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.136364 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.141384 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-2n4tf"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.141894 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.142340 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5xv2z"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.142775 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-5xv2z" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.147236 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-2n4tf" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.147543 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.147607 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.148031 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.148106 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9ws8"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.148560 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.148772 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.149066 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.158610 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.159873 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.161580 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.162106 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.196897 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.198574 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-wv7m2"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.198929 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.199048 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.199380 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.199516 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.199653 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.200061 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.200200 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.208353 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.208484 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.208629 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.208686 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.208708 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.208782 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.208857 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.209173 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.209221 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.209608 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.209721 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.210117 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.210270 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.210351 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.210391 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.210505 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.210508 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.210574 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.210596 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.210714 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.210837 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.210968 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211077 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211115 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211151 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211294 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211331 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211407 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211439 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211295 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211513 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211557 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211580 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211678 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211705 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211725 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211863 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211915 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.211937 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212025 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212033 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212063 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212028 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212109 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212161 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212165 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212177 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212191 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212301 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212381 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212563 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212691 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v4mm\" (UniqueName: \"kubernetes.io/projected/ec601e1d-d940-4bbb-9cce-83be390511ef-kube-api-access-5v4mm\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212725 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4m8vp\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212772 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212789 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-audit-policies\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212821 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212850 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212898 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212929 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/29babfee-4ec4-4828-90c6-caf8fb0647b2-metrics-tls\") pod \"ingress-operator-5b745b69d9-rzv4n\" (UID: \"29babfee-4ec4-4828-90c6-caf8fb0647b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.215868 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.216735 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.221306 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.221701 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.222398 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.212963 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29babfee-4ec4-4828-90c6-caf8fb0647b2-trusted-ca\") pod \"ingress-operator-5b745b69d9-rzv4n\" (UID: \"29babfee-4ec4-4828-90c6-caf8fb0647b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.222611 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-client-ca\") pod \"controller-manager-879f6c89f-4m8vp\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.222643 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2qkr\" (UniqueName: \"kubernetes.io/projected/1000b998-ffa1-40c5-8563-0aeb70d172a9-kube-api-access-c2qkr\") pod \"controller-manager-879f6c89f-4m8vp\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.222681 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.222715 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.222852 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.222992 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223041 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223073 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223197 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.222723 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223329 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223353 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223375 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223381 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8b44d028-0edc-4477-9f01-236c6d59d9e8-machine-approver-tls\") pod \"machine-approver-56656f9798-ktzsn\" (UID: \"8b44d028-0edc-4477-9f01-236c6d59d9e8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223408 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b44d028-0edc-4477-9f01-236c6d59d9e8-config\") pod \"machine-approver-56656f9798-ktzsn\" (UID: \"8b44d028-0edc-4477-9f01-236c6d59d9e8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223432 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1000b998-ffa1-40c5-8563-0aeb70d172a9-serving-cert\") pod \"controller-manager-879f6c89f-4m8vp\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223461 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223463 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l8vf\" (UniqueName: \"kubernetes.io/projected/8b44d028-0edc-4477-9f01-236c6d59d9e8-kube-api-access-9l8vf\") pod \"machine-approver-56656f9798-ktzsn\" (UID: \"8b44d028-0edc-4477-9f01-236c6d59d9e8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223549 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223562 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223568 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwvb7\" (UniqueName: \"kubernetes.io/projected/29babfee-4ec4-4828-90c6-caf8fb0647b2-kube-api-access-jwvb7\") pod \"ingress-operator-5b745b69d9-rzv4n\" (UID: \"29babfee-4ec4-4828-90c6-caf8fb0647b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223602 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223634 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223685 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/29babfee-4ec4-4828-90c6-caf8fb0647b2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-rzv4n\" (UID: \"29babfee-4ec4-4828-90c6-caf8fb0647b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223714 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-294s4\" (UniqueName: \"kubernetes.io/projected/6bc61e1c-d238-49ae-b344-429bf50959e5-kube-api-access-294s4\") pod \"cluster-samples-operator-665b6dd947-4rvqh\" (UID: \"6bc61e1c-d238-49ae-b344-429bf50959e5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4rvqh" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223771 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6bc61e1c-d238-49ae-b344-429bf50959e5-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4rvqh\" (UID: \"6bc61e1c-d238-49ae-b344-429bf50959e5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4rvqh" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223809 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223837 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223861 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8b44d028-0edc-4477-9f01-236c6d59d9e8-auth-proxy-config\") pod \"machine-approver-56656f9798-ktzsn\" (UID: \"8b44d028-0edc-4477-9f01-236c6d59d9e8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223895 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-config\") pod \"controller-manager-879f6c89f-4m8vp\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223925 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ec601e1d-d940-4bbb-9cce-83be390511ef-audit-dir\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.223950 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.224660 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.225104 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.225989 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.232102 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.233784 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.234769 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.255770 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.256054 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.257234 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.257441 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.266434 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.267014 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.268055 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.272666 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-2n4tf"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.274416 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.275075 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.275199 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-twcnw"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.275866 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twcnw" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.277176 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.279390 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jds5c"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.279836 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.279954 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.284139 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mjlld"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.285062 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.285558 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.286207 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.286454 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.286809 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.288330 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ws5nw"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.288852 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.289305 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.289306 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ws5nw" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.289362 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.290214 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.290235 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.291268 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.292099 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-h8mkp"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.293231 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ncx8v"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.293703 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.294203 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.294532 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ncx8v" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.294591 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.295847 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.296263 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.296774 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4xl77"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.297499 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xl77" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.297872 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536826-dgwn4"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.298664 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536826-dgwn4" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.299434 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.299436 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.302107 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.302755 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.304581 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9ws8"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.306516 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-qzwns"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.307302 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-qzwns" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.307786 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4m8vp"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.310663 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4rvqh"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.313474 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pbsmc"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.314544 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5xv2z"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.318888 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.319001 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.319924 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-twcnw"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.321114 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.322368 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-k4p69"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.323422 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.324606 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v4mm\" (UniqueName: \"kubernetes.io/projected/ec601e1d-d940-4bbb-9cce-83be390511ef-kube-api-access-5v4mm\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.324639 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4m8vp\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.324690 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-audit-policies\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.324711 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.324735 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.324760 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.324798 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/29babfee-4ec4-4828-90c6-caf8fb0647b2-metrics-tls\") pod \"ingress-operator-5b745b69d9-rzv4n\" (UID: \"29babfee-4ec4-4828-90c6-caf8fb0647b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.324836 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29babfee-4ec4-4828-90c6-caf8fb0647b2-trusted-ca\") pod \"ingress-operator-5b745b69d9-rzv4n\" (UID: \"29babfee-4ec4-4828-90c6-caf8fb0647b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.324856 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-client-ca\") pod \"controller-manager-879f6c89f-4m8vp\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.324876 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2qkr\" (UniqueName: \"kubernetes.io/projected/1000b998-ffa1-40c5-8563-0aeb70d172a9-kube-api-access-c2qkr\") pod \"controller-manager-879f6c89f-4m8vp\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.324896 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.324918 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.324942 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.324959 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8b44d028-0edc-4477-9f01-236c6d59d9e8-machine-approver-tls\") pod \"machine-approver-56656f9798-ktzsn\" (UID: \"8b44d028-0edc-4477-9f01-236c6d59d9e8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.324980 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b44d028-0edc-4477-9f01-236c6d59d9e8-config\") pod \"machine-approver-56656f9798-ktzsn\" (UID: \"8b44d028-0edc-4477-9f01-236c6d59d9e8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.325001 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1000b998-ffa1-40c5-8563-0aeb70d172a9-serving-cert\") pod \"controller-manager-879f6c89f-4m8vp\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.325024 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l8vf\" (UniqueName: \"kubernetes.io/projected/8b44d028-0edc-4477-9f01-236c6d59d9e8-kube-api-access-9l8vf\") pod \"machine-approver-56656f9798-ktzsn\" (UID: \"8b44d028-0edc-4477-9f01-236c6d59d9e8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.325045 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwvb7\" (UniqueName: \"kubernetes.io/projected/29babfee-4ec4-4828-90c6-caf8fb0647b2-kube-api-access-jwvb7\") pod \"ingress-operator-5b745b69d9-rzv4n\" (UID: \"29babfee-4ec4-4828-90c6-caf8fb0647b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.325065 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.325086 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.325118 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/29babfee-4ec4-4828-90c6-caf8fb0647b2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-rzv4n\" (UID: \"29babfee-4ec4-4828-90c6-caf8fb0647b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.325138 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-294s4\" (UniqueName: \"kubernetes.io/projected/6bc61e1c-d238-49ae-b344-429bf50959e5-kube-api-access-294s4\") pod \"cluster-samples-operator-665b6dd947-4rvqh\" (UID: \"6bc61e1c-d238-49ae-b344-429bf50959e5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4rvqh" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.325241 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6bc61e1c-d238-49ae-b344-429bf50959e5-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4rvqh\" (UID: \"6bc61e1c-d238-49ae-b344-429bf50959e5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4rvqh" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.325602 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.325648 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.325673 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8b44d028-0edc-4477-9f01-236c6d59d9e8-auth-proxy-config\") pod \"machine-approver-56656f9798-ktzsn\" (UID: \"8b44d028-0edc-4477-9f01-236c6d59d9e8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.325694 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-config\") pod \"controller-manager-879f6c89f-4m8vp\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.325717 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ec601e1d-d940-4bbb-9cce-83be390511ef-audit-dir\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.325738 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.326302 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pwdrv"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.326354 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.326366 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.327749 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b44d028-0edc-4477-9f01-236c6d59d9e8-config\") pod \"machine-approver-56656f9798-ktzsn\" (UID: \"8b44d028-0edc-4477-9f01-236c6d59d9e8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.327762 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4m8vp\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.328269 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-client-ca\") pod \"controller-manager-879f6c89f-4m8vp\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.328384 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-audit-policies\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.328793 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8b44d028-0edc-4477-9f01-236c6d59d9e8-auth-proxy-config\") pod \"machine-approver-56656f9798-ktzsn\" (UID: \"8b44d028-0edc-4477-9f01-236c6d59d9e8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.335676 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.336852 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ncx8v"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.337047 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.337343 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.337655 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7x2qp"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.337660 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.337716 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8b44d028-0edc-4477-9f01-236c6d59d9e8-machine-approver-tls\") pod \"machine-approver-56656f9798-ktzsn\" (UID: \"8b44d028-0edc-4477-9f01-236c6d59d9e8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.338215 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.338861 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.339403 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.340295 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.340332 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.340441 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ec601e1d-d940-4bbb-9cce-83be390511ef-audit-dir\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.341472 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.341551 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-config\") pod \"controller-manager-879f6c89f-4m8vp\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.342021 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.343337 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.345644 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6bc61e1c-d238-49ae-b344-429bf50959e5-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4rvqh\" (UID: \"6bc61e1c-d238-49ae-b344-429bf50959e5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4rvqh" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.348667 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9vmj5"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.349963 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.351073 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-6hpl2"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.351634 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1000b998-ffa1-40c5-8563-0aeb70d172a9-serving-cert\") pod \"controller-manager-879f6c89f-4m8vp\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.351966 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-6hpl2" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.352598 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.353362 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.354355 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536826-dgwn4"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.355585 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.356692 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-m9qmb"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.357361 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.357467 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-m9qmb" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.358121 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ndc8k"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.359352 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ws5nw"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.360060 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mjlld"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.361265 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.362402 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4xl77"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.363807 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-h8mkp"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.364928 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.365917 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-qzwns"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.367022 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.368011 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.369170 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-6hpl2"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.370244 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.371193 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-sbf5t"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.372104 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-sbf5t" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.372146 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-sbf5t"] Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.377301 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.397891 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.427976 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.438129 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.458279 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.477969 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.486524 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.486575 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.486603 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.486575 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.497919 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.517864 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.538546 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.557895 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.578221 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.591158 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/29babfee-4ec4-4828-90c6-caf8fb0647b2-metrics-tls\") pod \"ingress-operator-5b745b69d9-rzv4n\" (UID: \"29babfee-4ec4-4828-90c6-caf8fb0647b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.612207 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.619056 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29babfee-4ec4-4828-90c6-caf8fb0647b2-trusted-ca\") pod \"ingress-operator-5b745b69d9-rzv4n\" (UID: \"29babfee-4ec4-4828-90c6-caf8fb0647b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.619538 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.657757 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.677974 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.698941 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.718346 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.738637 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.757691 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.778053 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.818947 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.839030 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.858917 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.878724 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.898641 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.919142 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.938401 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.959001 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.979498 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 27 16:26:49 crc kubenswrapper[4814]: I0227 16:26:49.999645 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.019068 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.038832 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.057524 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.077645 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.098697 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.118514 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.138517 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.158116 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.179876 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.198313 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.218046 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.249411 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.260010 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.278404 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.296077 4814 request.go:700] Waited for 1.010488043s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.298118 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.318477 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.339395 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.359845 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.378230 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.397899 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.418199 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.437899 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.459054 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.478588 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.497820 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.518725 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.537617 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.558086 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.578950 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.598363 4814 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.618747 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.639333 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.660047 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.679008 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.698993 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.718052 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.740114 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.758617 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.778699 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.799794 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.818548 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.839395 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.857859 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.878974 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.898121 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.917993 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.938739 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.959708 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 27 16:26:50 crc kubenswrapper[4814]: I0227 16:26:50.978329 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.027109 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v4mm\" (UniqueName: \"kubernetes.io/projected/ec601e1d-d940-4bbb-9cce-83be390511ef-kube-api-access-5v4mm\") pod \"oauth-openshift-558db77b4-k4p69\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.036840 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2qkr\" (UniqueName: \"kubernetes.io/projected/1000b998-ffa1-40c5-8563-0aeb70d172a9-kube-api-access-c2qkr\") pod \"controller-manager-879f6c89f-4m8vp\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.068495 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.070658 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-294s4\" (UniqueName: \"kubernetes.io/projected/6bc61e1c-d238-49ae-b344-429bf50959e5-kube-api-access-294s4\") pod \"cluster-samples-operator-665b6dd947-4rvqh\" (UID: \"6bc61e1c-d238-49ae-b344-429bf50959e5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4rvqh" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.086294 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l8vf\" (UniqueName: \"kubernetes.io/projected/8b44d028-0edc-4477-9f01-236c6d59d9e8-kube-api-access-9l8vf\") pod \"machine-approver-56656f9798-ktzsn\" (UID: \"8b44d028-0edc-4477-9f01-236c6d59d9e8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.089028 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.111638 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwvb7\" (UniqueName: \"kubernetes.io/projected/29babfee-4ec4-4828-90c6-caf8fb0647b2-kube-api-access-jwvb7\") pod \"ingress-operator-5b745b69d9-rzv4n\" (UID: \"29babfee-4ec4-4828-90c6-caf8fb0647b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.142460 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.142598 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.148170 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/29babfee-4ec4-4828-90c6-caf8fb0647b2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-rzv4n\" (UID: \"29babfee-4ec4-4828-90c6-caf8fb0647b2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.159818 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.178612 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.198521 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.198809 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.217524 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.237943 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.242842 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.260395 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.280324 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.296669 4814 request.go:700] Waited for 1.924142873s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/secrets?fieldSelector=metadata.name%3Ddns-default-metrics-tls&limit=500&resourceVersion=0 Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.299775 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.319188 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.327683 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4rvqh" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.340199 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.361001 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.378926 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.398144 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.418439 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.453804 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-config\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.453867 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/008442d1-8ae3-4669-945d-d7f3389855a2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-t6c54\" (UID: \"008442d1-8ae3-4669-945d-d7f3389855a2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.453899 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46c9f2b1-15f1-418b-a086-5ea9972185a0-config\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.453926 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-xx5hs\" (UID: \"d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.453953 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cdc3c78e-177f-4fcd-ab8d-c333497848aa-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-nh2sn\" (UID: \"cdc3c78e-177f-4fcd-ab8d-c333497848aa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.453976 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e245069f-c15d-4ad4-9709-6bec4a52c8c8-config\") pod \"authentication-operator-69f744f599-jds5c\" (UID: \"e245069f-c15d-4ad4-9709-6bec4a52c8c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454000 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlsxd\" (UniqueName: \"kubernetes.io/projected/46c9f2b1-15f1-418b-a086-5ea9972185a0-kube-api-access-mlsxd\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454025 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b4a4d02-59a6-4f74-a4cc-2154d1115361-ca-trust-extracted\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454052 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b8fa45e-2136-44a1-9a50-f936b5c22417-config\") pod \"route-controller-manager-6576b87f9c-gmfbn\" (UID: \"2b8fa45e-2136-44a1-9a50-f936b5c22417\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454076 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-image-import-ca\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454117 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db53a48c-000d-4e92-929e-6c6bf561119a-serving-cert\") pod \"console-operator-58897d9998-5xv2z\" (UID: \"db53a48c-000d-4e92-929e-6c6bf561119a\") " pod="openshift-console-operator/console-operator-58897d9998-5xv2z" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454142 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-serving-cert\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454171 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454195 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-oauth-serving-cert\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454218 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fc590a1b-ba51-421c-a089-fb0def054fb2-metrics-tls\") pod \"dns-operator-744455d44c-pwdrv\" (UID: \"fc590a1b-ba51-421c-a089-fb0def054fb2\") " pod="openshift-dns-operator/dns-operator-744455d44c-pwdrv" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454239 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-etcd-client\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454281 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/008442d1-8ae3-4669-945d-d7f3389855a2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-t6c54\" (UID: \"008442d1-8ae3-4669-945d-d7f3389855a2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454306 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-audit-dir\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454338 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7cqz\" (UniqueName: \"kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-kube-api-access-m7cqz\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454364 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l92p\" (UniqueName: \"kubernetes.io/projected/fc590a1b-ba51-421c-a089-fb0def054fb2-kube-api-access-9l92p\") pod \"dns-operator-744455d44c-pwdrv\" (UID: \"fc590a1b-ba51-421c-a089-fb0def054fb2\") " pod="openshift-dns-operator/dns-operator-744455d44c-pwdrv" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454391 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc099c2d-2785-4509-bd8b-29469341f56d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ndc8k\" (UID: \"bc099c2d-2785-4509-bd8b-29469341f56d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454440 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e245069f-c15d-4ad4-9709-6bec4a52c8c8-serving-cert\") pod \"authentication-operator-69f744f599-jds5c\" (UID: \"e245069f-c15d-4ad4-9709-6bec4a52c8c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454475 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq2xf\" (UniqueName: \"kubernetes.io/projected/bc099c2d-2785-4509-bd8b-29469341f56d-kube-api-access-cq2xf\") pod \"machine-api-operator-5694c8668f-ndc8k\" (UID: \"bc099c2d-2785-4509-bd8b-29469341f56d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454499 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djkd5\" (UniqueName: \"kubernetes.io/projected/e245069f-c15d-4ad4-9709-6bec4a52c8c8-kube-api-access-djkd5\") pod \"authentication-operator-69f744f599-jds5c\" (UID: \"e245069f-c15d-4ad4-9709-6bec4a52c8c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454521 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/46c9f2b1-15f1-418b-a086-5ea9972185a0-etcd-ca\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454546 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b8fa45e-2136-44a1-9a50-f936b5c22417-serving-cert\") pod \"route-controller-manager-6576b87f9c-gmfbn\" (UID: \"2b8fa45e-2136-44a1-9a50-f936b5c22417\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454568 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-config\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454597 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cdc3c78e-177f-4fcd-ab8d-c333497848aa-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-nh2sn\" (UID: \"cdc3c78e-177f-4fcd-ab8d-c333497848aa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454661 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bc099c2d-2785-4509-bd8b-29469341f56d-images\") pod \"machine-api-operator-5694c8668f-ndc8k\" (UID: \"bc099c2d-2785-4509-bd8b-29469341f56d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454685 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b4a4d02-59a6-4f74-a4cc-2154d1115361-trusted-ca\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454711 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-xx5hs\" (UID: \"d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454734 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc099c2d-2785-4509-bd8b-29469341f56d-config\") pod \"machine-api-operator-5694c8668f-ndc8k\" (UID: \"bc099c2d-2785-4509-bd8b-29469341f56d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454754 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-audit-policies\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454778 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv48m\" (UniqueName: \"kubernetes.io/projected/04649429-b45f-41ff-8777-4e0b321664b9-kube-api-access-wv48m\") pod \"openshift-config-operator-7777fb866f-qnsbd\" (UID: \"04649429-b45f-41ff-8777-4e0b321664b9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454800 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-etcd-client\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454833 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-serving-cert\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454860 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e245069f-c15d-4ad4-9709-6bec4a52c8c8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jds5c\" (UID: \"e245069f-c15d-4ad4-9709-6bec4a52c8c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454884 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b8fa45e-2136-44a1-9a50-f936b5c22417-client-ca\") pod \"route-controller-manager-6576b87f9c-gmfbn\" (UID: \"2b8fa45e-2136-44a1-9a50-f936b5c22417\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454935 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b4a4d02-59a6-4f74-a4cc-2154d1115361-installation-pull-secrets\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454962 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e245069f-c15d-4ad4-9709-6bec4a52c8c8-service-ca-bundle\") pod \"authentication-operator-69f744f599-jds5c\" (UID: \"e245069f-c15d-4ad4-9709-6bec4a52c8c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.454986 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-encryption-config\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.455011 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/46c9f2b1-15f1-418b-a086-5ea9972185a0-etcd-service-ca\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.455033 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/46c9f2b1-15f1-418b-a086-5ea9972185a0-etcd-client\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.456853 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-serving-cert\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.456956 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b4a4d02-59a6-4f74-a4cc-2154d1115361-registry-certificates\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457003 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/cdc3c78e-177f-4fcd-ab8d-c333497848aa-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-nh2sn\" (UID: \"cdc3c78e-177f-4fcd-ab8d-c333497848aa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457040 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8tlp\" (UniqueName: \"kubernetes.io/projected/d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c-kube-api-access-t8tlp\") pod \"openshift-controller-manager-operator-756b6f6bc6-xx5hs\" (UID: \"d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457106 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457141 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-audit\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457176 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54klp\" (UniqueName: \"kubernetes.io/projected/cdc3c78e-177f-4fcd-ab8d-c333497848aa-kube-api-access-54klp\") pod \"cluster-image-registry-operator-dc59b4c8b-nh2sn\" (UID: \"cdc3c78e-177f-4fcd-ab8d-c333497848aa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457211 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfmj9\" (UniqueName: \"kubernetes.io/projected/008442d1-8ae3-4669-945d-d7f3389855a2-kube-api-access-lfmj9\") pod \"openshift-apiserver-operator-796bbdcf4f-t6c54\" (UID: \"008442d1-8ae3-4669-945d-d7f3389855a2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457247 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db53a48c-000d-4e92-929e-6c6bf561119a-trusted-ca\") pod \"console-operator-58897d9998-5xv2z\" (UID: \"db53a48c-000d-4e92-929e-6c6bf561119a\") " pod="openshift-console-operator/console-operator-58897d9998-5xv2z" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457351 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-oauth-config\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457385 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457421 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457461 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-service-ca\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457511 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db53a48c-000d-4e92-929e-6c6bf561119a-config\") pod \"console-operator-58897d9998-5xv2z\" (UID: \"db53a48c-000d-4e92-929e-6c6bf561119a\") " pod="openshift-console-operator/console-operator-58897d9998-5xv2z" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457564 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c9r4\" (UniqueName: \"kubernetes.io/projected/2b8fa45e-2136-44a1-9a50-f936b5c22417-kube-api-access-4c9r4\") pod \"route-controller-manager-6576b87f9c-gmfbn\" (UID: \"2b8fa45e-2136-44a1-9a50-f936b5c22417\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457599 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmknd\" (UniqueName: \"kubernetes.io/projected/f487b7f4-c24b-431e-8868-a0285c8ed7bc-kube-api-access-qmknd\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457631 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04649429-b45f-41ff-8777-4e0b321664b9-serving-cert\") pod \"openshift-config-operator-7777fb866f-qnsbd\" (UID: \"04649429-b45f-41ff-8777-4e0b321664b9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457660 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-encryption-config\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457691 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8dbk\" (UniqueName: \"kubernetes.io/projected/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-kube-api-access-d8dbk\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457744 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-bound-sa-token\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457780 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph5x9\" (UniqueName: \"kubernetes.io/projected/db53a48c-000d-4e92-929e-6c6bf561119a-kube-api-access-ph5x9\") pod \"console-operator-58897d9998-5xv2z\" (UID: \"db53a48c-000d-4e92-929e-6c6bf561119a\") " pod="openshift-console-operator/console-operator-58897d9998-5xv2z" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457812 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46c9f2b1-15f1-418b-a086-5ea9972185a0-serving-cert\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457861 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-etcd-serving-ca\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457890 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-audit-dir\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457927 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-registry-tls\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457958 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-trusted-ca-bundle\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.457987 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/04649429-b45f-41ff-8777-4e0b321664b9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-qnsbd\" (UID: \"04649429-b45f-41ff-8777-4e0b321664b9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.458017 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-node-pullsecrets\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.458054 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j79wm\" (UniqueName: \"kubernetes.io/projected/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-kube-api-access-j79wm\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.463692 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4g45\" (UniqueName: \"kubernetes.io/projected/98112972-5121-4d93-8490-5156a1a2f3ca-kube-api-access-m4g45\") pod \"downloads-7954f5f757-2n4tf\" (UID: \"98112972-5121-4d93-8490-5156a1a2f3ca\") " pod="openshift-console/downloads-7954f5f757-2n4tf" Feb 27 16:26:51 crc kubenswrapper[4814]: E0227 16:26:51.464875 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:51.964845528 +0000 UTC m=+224.417470378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.564871 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565161 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cdc3c78e-177f-4fcd-ab8d-c333497848aa-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-nh2sn\" (UID: \"cdc3c78e-177f-4fcd-ab8d-c333497848aa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565187 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7dc9\" (UniqueName: \"kubernetes.io/projected/5768bed9-9187-4bef-ae5e-bc003793d741-kube-api-access-t7dc9\") pod \"package-server-manager-789f6589d5-6c5hj\" (UID: \"5768bed9-9187-4bef-ae5e-bc003793d741\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565204 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3b269e43-ebd6-4f71-8963-669c60763efa-proxy-tls\") pod \"machine-config-controller-84d6567774-g4vb2\" (UID: \"3b269e43-ebd6-4f71-8963-669c60763efa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565224 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/0f52f92d-a4bd-41b3-88dd-579a355f6e30-plugins-dir\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565241 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3e0b793d-24a2-4c73-9357-fe193c32f71b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-w4f4k\" (UID: \"3e0b793d-24a2-4c73-9357-fe193c32f71b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565509 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-xx5hs\" (UID: \"d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565538 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc099c2d-2785-4509-bd8b-29469341f56d-config\") pod \"machine-api-operator-5694c8668f-ndc8k\" (UID: \"bc099c2d-2785-4509-bd8b-29469341f56d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565558 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9b78e04c-f222-4aae-be52-85abd4ecf391-metrics-tls\") pod \"dns-default-sbf5t\" (UID: \"9b78e04c-f222-4aae-be52-85abd4ecf391\") " pod="openshift-dns/dns-default-sbf5t" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565585 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv48m\" (UniqueName: \"kubernetes.io/projected/04649429-b45f-41ff-8777-4e0b321664b9-kube-api-access-wv48m\") pod \"openshift-config-operator-7777fb866f-qnsbd\" (UID: \"04649429-b45f-41ff-8777-4e0b321664b9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565603 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b8fa45e-2136-44a1-9a50-f936b5c22417-client-ca\") pod \"route-controller-manager-6576b87f9c-gmfbn\" (UID: \"2b8fa45e-2136-44a1-9a50-f936b5c22417\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565619 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-etcd-client\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565643 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/057f89ed-0870-4017-8a0c-a4808d3fb86b-apiservice-cert\") pod \"packageserver-d55dfcdfc-8tt9s\" (UID: \"057f89ed-0870-4017-8a0c-a4808d3fb86b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565665 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e245069f-c15d-4ad4-9709-6bec4a52c8c8-service-ca-bundle\") pod \"authentication-operator-69f744f599-jds5c\" (UID: \"e245069f-c15d-4ad4-9709-6bec4a52c8c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565683 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97d5e809-690b-44bc-9df0-1af46832e149-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lmzk2\" (UID: \"97d5e809-690b-44bc-9df0-1af46832e149\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565701 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/46c9f2b1-15f1-418b-a086-5ea9972185a0-etcd-service-ca\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565738 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/46c9f2b1-15f1-418b-a086-5ea9972185a0-etcd-client\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565755 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3e0b793d-24a2-4c73-9357-fe193c32f71b-images\") pod \"machine-config-operator-74547568cd-w4f4k\" (UID: \"3e0b793d-24a2-4c73-9357-fe193c32f71b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" Feb 27 16:26:51 crc kubenswrapper[4814]: E0227 16:26:51.565793 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:52.065756333 +0000 UTC m=+224.518381193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565883 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-serving-cert\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565926 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/cdc3c78e-177f-4fcd-ab8d-c333497848aa-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-nh2sn\" (UID: \"cdc3c78e-177f-4fcd-ab8d-c333497848aa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.565991 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566033 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5768bed9-9187-4bef-ae5e-bc003793d741-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-6c5hj\" (UID: \"5768bed9-9187-4bef-ae5e-bc003793d741\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566093 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpbj9\" (UniqueName: \"kubernetes.io/projected/9541f10f-b16a-4dbe-8de2-fb4f381fdb94-kube-api-access-gpbj9\") pod \"kube-storage-version-migrator-operator-b67b599dd-gnk2x\" (UID: \"9541f10f-b16a-4dbe-8de2-fb4f381fdb94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566126 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32d7bef-558f-4dfd-9260-408cb47fc461-config\") pod \"kube-apiserver-operator-766d6c64bb-rs9mx\" (UID: \"d32d7bef-558f-4dfd-9260-408cb47fc461\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566158 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db53a48c-000d-4e92-929e-6c6bf561119a-config\") pod \"console-operator-58897d9998-5xv2z\" (UID: \"db53a48c-000d-4e92-929e-6c6bf561119a\") " pod="openshift-console-operator/console-operator-58897d9998-5xv2z" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566196 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbhfj\" (UniqueName: \"kubernetes.io/projected/9b78e04c-f222-4aae-be52-85abd4ecf391-kube-api-access-cbhfj\") pod \"dns-default-sbf5t\" (UID: \"9b78e04c-f222-4aae-be52-85abd4ecf391\") " pod="openshift-dns/dns-default-sbf5t" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566289 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c9r4\" (UniqueName: \"kubernetes.io/projected/2b8fa45e-2136-44a1-9a50-f936b5c22417-kube-api-access-4c9r4\") pod \"route-controller-manager-6576b87f9c-gmfbn\" (UID: \"2b8fa45e-2136-44a1-9a50-f936b5c22417\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566347 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx4gm\" (UniqueName: \"kubernetes.io/projected/3b269e43-ebd6-4f71-8963-669c60763efa-kube-api-access-zx4gm\") pod \"machine-config-controller-84d6567774-g4vb2\" (UID: \"3b269e43-ebd6-4f71-8963-669c60763efa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566381 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/13240f27-5cc2-4e90-9992-24f6beca4212-default-certificate\") pod \"router-default-5444994796-wv7m2\" (UID: \"13240f27-5cc2-4e90-9992-24f6beca4212\") " pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566413 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6401ae31-5292-4d4f-8d39-a5682c0d38ee-srv-cert\") pod \"catalog-operator-68c6474976-tzdzl\" (UID: \"6401ae31-5292-4d4f-8d39-a5682c0d38ee\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566446 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmknd\" (UniqueName: \"kubernetes.io/projected/f487b7f4-c24b-431e-8868-a0285c8ed7bc-kube-api-access-qmknd\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566466 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-xx5hs\" (UID: \"d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566479 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-encryption-config\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566549 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-audit-dir\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566580 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55b73477-5f70-4dcc-a9aa-3e22ac4efb54-cert\") pod \"ingress-canary-6hpl2\" (UID: \"55b73477-5f70-4dcc-a9aa-3e22ac4efb54\") " pod="openshift-ingress-canary/ingress-canary-6hpl2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566614 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13240f27-5cc2-4e90-9992-24f6beca4212-metrics-certs\") pod \"router-default-5444994796-wv7m2\" (UID: \"13240f27-5cc2-4e90-9992-24f6beca4212\") " pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566652 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-trusted-ca-bundle\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.566686 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/04649429-b45f-41ff-8777-4e0b321664b9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-qnsbd\" (UID: \"04649429-b45f-41ff-8777-4e0b321664b9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.567133 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc099c2d-2785-4509-bd8b-29469341f56d-config\") pod \"machine-api-operator-5694c8668f-ndc8k\" (UID: \"bc099c2d-2785-4509-bd8b-29469341f56d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" Feb 27 16:26:51 crc kubenswrapper[4814]: E0227 16:26:51.567329 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:52.067302956 +0000 UTC m=+224.519927786 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.567788 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e245069f-c15d-4ad4-9709-6bec4a52c8c8-service-ca-bundle\") pod \"authentication-operator-69f744f599-jds5c\" (UID: \"e245069f-c15d-4ad4-9709-6bec4a52c8c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.567911 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/04649429-b45f-41ff-8777-4e0b321664b9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-qnsbd\" (UID: \"04649429-b45f-41ff-8777-4e0b321664b9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.567929 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-audit-dir\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.568027 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-node-pullsecrets\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.568067 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-node-pullsecrets\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.568103 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3b269e43-ebd6-4f71-8963-669c60763efa-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-g4vb2\" (UID: \"3b269e43-ebd6-4f71-8963-669c60763efa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.568327 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggb4k\" (UniqueName: \"kubernetes.io/projected/d87f2169-8a70-42de-aaf5-7728ff95fa50-kube-api-access-ggb4k\") pod \"collect-profiles-29536815-pqwsb\" (UID: \"d87f2169-8a70-42de-aaf5-7728ff95fa50\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.568516 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/057f89ed-0870-4017-8a0c-a4808d3fb86b-tmpfs\") pod \"packageserver-d55dfcdfc-8tt9s\" (UID: \"057f89ed-0870-4017-8a0c-a4808d3fb86b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.568700 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13240f27-5cc2-4e90-9992-24f6beca4212-service-ca-bundle\") pod \"router-default-5444994796-wv7m2\" (UID: \"13240f27-5cc2-4e90-9992-24f6beca4212\") " pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.568742 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-config\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.568767 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/008442d1-8ae3-4669-945d-d7f3389855a2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-t6c54\" (UID: \"008442d1-8ae3-4669-945d-d7f3389855a2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.568826 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cdc3c78e-177f-4fcd-ab8d-c333497848aa-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-nh2sn\" (UID: \"cdc3c78e-177f-4fcd-ab8d-c333497848aa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.568831 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/46c9f2b1-15f1-418b-a086-5ea9972185a0-etcd-service-ca\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.568851 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlsxd\" (UniqueName: \"kubernetes.io/projected/46c9f2b1-15f1-418b-a086-5ea9972185a0-kube-api-access-mlsxd\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.568897 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nstsm\" (UniqueName: \"kubernetes.io/projected/634e6fbd-87f9-4205-b01d-90be6eb4a015-kube-api-access-nstsm\") pod \"migrator-59844c95c7-twcnw\" (UID: \"634e6fbd-87f9-4205-b01d-90be6eb4a015\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twcnw" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.568919 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db53a48c-000d-4e92-929e-6c6bf561119a-config\") pod \"console-operator-58897d9998-5xv2z\" (UID: \"db53a48c-000d-4e92-929e-6c6bf561119a\") " pod="openshift-console-operator/console-operator-58897d9998-5xv2z" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.569097 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b8fa45e-2136-44a1-9a50-f936b5c22417-config\") pod \"route-controller-manager-6576b87f9c-gmfbn\" (UID: \"2b8fa45e-2136-44a1-9a50-f936b5c22417\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.569162 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b4a4d02-59a6-4f74-a4cc-2154d1115361-ca-trust-extracted\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.569473 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ba54dbcb-efbd-467b-aa4e-994fde5e1c14-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ncx8v\" (UID: \"ba54dbcb-efbd-467b-aa4e-994fde5e1c14\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ncx8v" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.569511 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/0f52f92d-a4bd-41b3-88dd-579a355f6e30-mountpoint-dir\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.569643 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/008442d1-8ae3-4669-945d-d7f3389855a2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-t6c54\" (UID: \"008442d1-8ae3-4669-945d-d7f3389855a2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.569846 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-config\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.569983 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b8fa45e-2136-44a1-9a50-f936b5c22417-client-ca\") pod \"route-controller-manager-6576b87f9c-gmfbn\" (UID: \"2b8fa45e-2136-44a1-9a50-f936b5c22417\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570039 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w6zl\" (UniqueName: \"kubernetes.io/projected/95913cb4-39f1-44c0-ac49-0a2d51047679-kube-api-access-8w6zl\") pod \"auto-csr-approver-29536826-dgwn4\" (UID: \"95913cb4-39f1-44c0-ac49-0a2d51047679\") " pod="openshift-infra/auto-csr-approver-29536826-dgwn4" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570049 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b4a4d02-59a6-4f74-a4cc-2154d1115361-ca-trust-extracted\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570115 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9541f10f-b16a-4dbe-8de2-fb4f381fdb94-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gnk2x\" (UID: \"9541f10f-b16a-4dbe-8de2-fb4f381fdb94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570167 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48wrk\" (UniqueName: \"kubernetes.io/projected/8676c139-d437-4f7a-9234-de7c8a1b8dd1-kube-api-access-48wrk\") pod \"service-ca-9c57cc56f-qzwns\" (UID: \"8676c139-d437-4f7a-9234-de7c8a1b8dd1\") " pod="openshift-service-ca/service-ca-9c57cc56f-qzwns" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570196 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fc590a1b-ba51-421c-a089-fb0def054fb2-metrics-tls\") pod \"dns-operator-744455d44c-pwdrv\" (UID: \"fc590a1b-ba51-421c-a089-fb0def054fb2\") " pod="openshift-dns-operator/dns-operator-744455d44c-pwdrv" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570218 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570279 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d367fc6a-d4c4-4e63-81da-c547e07eea8d-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhq9n\" (UID: \"d367fc6a-d4c4-4e63-81da-c547e07eea8d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570299 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d87f2169-8a70-42de-aaf5-7728ff95fa50-config-volume\") pod \"collect-profiles-29536815-pqwsb\" (UID: \"d87f2169-8a70-42de-aaf5-7728ff95fa50\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570390 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/008442d1-8ae3-4669-945d-d7f3389855a2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-t6c54\" (UID: \"008442d1-8ae3-4669-945d-d7f3389855a2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570418 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-audit-dir\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570491 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d87f2169-8a70-42de-aaf5-7728ff95fa50-secret-volume\") pod \"collect-profiles-29536815-pqwsb\" (UID: \"d87f2169-8a70-42de-aaf5-7728ff95fa50\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570518 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfa49cf2-dc52-4c27-8486-ec1efec61744-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mjlld\" (UID: \"dfa49cf2-dc52-4c27-8486-ec1efec61744\") " pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570540 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/13240f27-5cc2-4e90-9992-24f6beca4212-stats-auth\") pod \"router-default-5444994796-wv7m2\" (UID: \"13240f27-5cc2-4e90-9992-24f6beca4212\") " pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570576 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr6lk\" (UniqueName: \"kubernetes.io/projected/13240f27-5cc2-4e90-9992-24f6beca4212-kube-api-access-lr6lk\") pod \"router-default-5444994796-wv7m2\" (UID: \"13240f27-5cc2-4e90-9992-24f6beca4212\") " pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570623 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7cqz\" (UniqueName: \"kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-kube-api-access-m7cqz\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570694 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq2xf\" (UniqueName: \"kubernetes.io/projected/bc099c2d-2785-4509-bd8b-29469341f56d-kube-api-access-cq2xf\") pod \"machine-api-operator-5694c8668f-ndc8k\" (UID: \"bc099c2d-2785-4509-bd8b-29469341f56d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570744 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d367fc6a-d4c4-4e63-81da-c547e07eea8d-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhq9n\" (UID: \"d367fc6a-d4c4-4e63-81da-c547e07eea8d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570804 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/0f52f92d-a4bd-41b3-88dd-579a355f6e30-socket-dir\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570837 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djkd5\" (UniqueName: \"kubernetes.io/projected/e245069f-c15d-4ad4-9709-6bec4a52c8c8-kube-api-access-djkd5\") pod \"authentication-operator-69f744f599-jds5c\" (UID: \"e245069f-c15d-4ad4-9709-6bec4a52c8c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570854 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b8fa45e-2136-44a1-9a50-f936b5c22417-serving-cert\") pod \"route-controller-manager-6576b87f9c-gmfbn\" (UID: \"2b8fa45e-2136-44a1-9a50-f936b5c22417\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570871 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-config\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570892 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3dafc1d6-a249-4d30-ad93-f065110076c8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ws5nw\" (UID: \"3dafc1d6-a249-4d30-ad93-f065110076c8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ws5nw" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570914 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8676c139-d437-4f7a-9234-de7c8a1b8dd1-signing-key\") pod \"service-ca-9c57cc56f-qzwns\" (UID: \"8676c139-d437-4f7a-9234-de7c8a1b8dd1\") " pod="openshift-service-ca/service-ca-9c57cc56f-qzwns" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570933 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld5b5\" (UniqueName: \"kubernetes.io/projected/55b73477-5f70-4dcc-a9aa-3e22ac4efb54-kube-api-access-ld5b5\") pod \"ingress-canary-6hpl2\" (UID: \"55b73477-5f70-4dcc-a9aa-3e22ac4efb54\") " pod="openshift-ingress-canary/ingress-canary-6hpl2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570949 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgcgk\" (UniqueName: \"kubernetes.io/projected/ba54dbcb-efbd-467b-aa4e-994fde5e1c14-kube-api-access-wgcgk\") pod \"multus-admission-controller-857f4d67dd-ncx8v\" (UID: \"ba54dbcb-efbd-467b-aa4e-994fde5e1c14\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ncx8v" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570971 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bc099c2d-2785-4509-bd8b-29469341f56d-images\") pod \"machine-api-operator-5694c8668f-ndc8k\" (UID: \"bc099c2d-2785-4509-bd8b-29469341f56d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.570997 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b8fa45e-2136-44a1-9a50-f936b5c22417-config\") pod \"route-controller-manager-6576b87f9c-gmfbn\" (UID: \"2b8fa45e-2136-44a1-9a50-f936b5c22417\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.571012 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b4a4d02-59a6-4f74-a4cc-2154d1115361-trusted-ca\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.571175 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-audit-policies\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.571212 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8676c139-d437-4f7a-9234-de7c8a1b8dd1-signing-cabundle\") pod \"service-ca-9c57cc56f-qzwns\" (UID: \"8676c139-d437-4f7a-9234-de7c8a1b8dd1\") " pod="openshift-service-ca/service-ca-9c57cc56f-qzwns" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.571249 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-trusted-ca-bundle\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.571317 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e245069f-c15d-4ad4-9709-6bec4a52c8c8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jds5c\" (UID: \"e245069f-c15d-4ad4-9709-6bec4a52c8c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.571351 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-serving-cert\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.571388 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-audit-dir\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.571394 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-encryption-config\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.571453 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b4a4d02-59a6-4f74-a4cc-2154d1115361-installation-pull-secrets\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.571492 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7kwh\" (UniqueName: \"kubernetes.io/projected/1a357d8b-0e0e-4ae0-9365-4c4d96316ae2-kube-api-access-p7kwh\") pod \"service-ca-operator-777779d784-4xl77\" (UID: \"1a357d8b-0e0e-4ae0-9365-4c4d96316ae2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xl77" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.572010 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.572997 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b4a4d02-59a6-4f74-a4cc-2154d1115361-trusted-ca\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.573901 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-audit-policies\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.574877 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cdc3c78e-177f-4fcd-ab8d-c333497848aa-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-nh2sn\" (UID: \"cdc3c78e-177f-4fcd-ab8d-c333497848aa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.575245 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/46c9f2b1-15f1-418b-a086-5ea9972185a0-etcd-client\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.575373 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e245069f-c15d-4ad4-9709-6bec4a52c8c8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jds5c\" (UID: \"e245069f-c15d-4ad4-9709-6bec4a52c8c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.575457 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b4a4d02-59a6-4f74-a4cc-2154d1115361-registry-certificates\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.575483 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8tlp\" (UniqueName: \"kubernetes.io/projected/d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c-kube-api-access-t8tlp\") pod \"openshift-controller-manager-operator-756b6f6bc6-xx5hs\" (UID: \"d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.575522 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97d5e809-690b-44bc-9df0-1af46832e149-config\") pod \"kube-controller-manager-operator-78b949d7b-lmzk2\" (UID: \"97d5e809-690b-44bc-9df0-1af46832e149\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.575802 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-config\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.576667 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b4a4d02-59a6-4f74-a4cc-2154d1115361-registry-certificates\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.576867 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d32d7bef-558f-4dfd-9260-408cb47fc461-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rs9mx\" (UID: \"d32d7bef-558f-4dfd-9260-408cb47fc461\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.576901 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/0f52f92d-a4bd-41b3-88dd-579a355f6e30-registration-dir\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.576931 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3e0b793d-24a2-4c73-9357-fe193c32f71b-proxy-tls\") pod \"machine-config-operator-74547568cd-w4f4k\" (UID: \"3e0b793d-24a2-4c73-9357-fe193c32f71b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.576980 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-audit\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577041 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54klp\" (UniqueName: \"kubernetes.io/projected/cdc3c78e-177f-4fcd-ab8d-c333497848aa-kube-api-access-54klp\") pod \"cluster-image-registry-operator-dc59b4c8b-nh2sn\" (UID: \"cdc3c78e-177f-4fcd-ab8d-c333497848aa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577066 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfmj9\" (UniqueName: \"kubernetes.io/projected/008442d1-8ae3-4669-945d-d7f3389855a2-kube-api-access-lfmj9\") pod \"openshift-apiserver-operator-796bbdcf4f-t6c54\" (UID: \"008442d1-8ae3-4669-945d-d7f3389855a2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577093 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32d7bef-558f-4dfd-9260-408cb47fc461-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rs9mx\" (UID: \"d32d7bef-558f-4dfd-9260-408cb47fc461\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577136 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db53a48c-000d-4e92-929e-6c6bf561119a-trusted-ca\") pod \"console-operator-58897d9998-5xv2z\" (UID: \"db53a48c-000d-4e92-929e-6c6bf561119a\") " pod="openshift-console-operator/console-operator-58897d9998-5xv2z" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577218 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-oauth-config\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577246 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577327 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-service-ca\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577348 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577381 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsmlh\" (UniqueName: \"kubernetes.io/projected/057f89ed-0870-4017-8a0c-a4808d3fb86b-kube-api-access-qsmlh\") pod \"packageserver-d55dfcdfc-8tt9s\" (UID: \"057f89ed-0870-4017-8a0c-a4808d3fb86b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577379 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bc099c2d-2785-4509-bd8b-29469341f56d-images\") pod \"machine-api-operator-5694c8668f-ndc8k\" (UID: \"bc099c2d-2785-4509-bd8b-29469341f56d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577410 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a357d8b-0e0e-4ae0-9365-4c4d96316ae2-config\") pod \"service-ca-operator-777779d784-4xl77\" (UID: \"1a357d8b-0e0e-4ae0-9365-4c4d96316ae2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xl77" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577530 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8bc5b690-2e68-4d62-a0cf-349326b8e5e4-srv-cert\") pod \"olm-operator-6b444d44fb-hl9wj\" (UID: \"8bc5b690-2e68-4d62-a0cf-349326b8e5e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577561 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04649429-b45f-41ff-8777-4e0b321664b9-serving-cert\") pod \"openshift-config-operator-7777fb866f-qnsbd\" (UID: \"04649429-b45f-41ff-8777-4e0b321664b9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577588 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8dbk\" (UniqueName: \"kubernetes.io/projected/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-kube-api-access-d8dbk\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577616 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46c9f2b1-15f1-418b-a086-5ea9972185a0-serving-cert\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577646 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmsmw\" (UniqueName: \"kubernetes.io/projected/dfa49cf2-dc52-4c27-8486-ec1efec61744-kube-api-access-qmsmw\") pod \"marketplace-operator-79b997595-mjlld\" (UID: \"dfa49cf2-dc52-4c27-8486-ec1efec61744\") " pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577678 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-bound-sa-token\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577701 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph5x9\" (UniqueName: \"kubernetes.io/projected/db53a48c-000d-4e92-929e-6c6bf561119a-kube-api-access-ph5x9\") pod \"console-operator-58897d9998-5xv2z\" (UID: \"db53a48c-000d-4e92-929e-6c6bf561119a\") " pod="openshift-console-operator/console-operator-58897d9998-5xv2z" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577750 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-etcd-serving-ca\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577789 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt2g7\" (UniqueName: \"kubernetes.io/projected/8bc5b690-2e68-4d62-a0cf-349326b8e5e4-kube-api-access-qt2g7\") pod \"olm-operator-6b444d44fb-hl9wj\" (UID: \"8bc5b690-2e68-4d62-a0cf-349326b8e5e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577813 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-registry-tls\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.577832 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j79wm\" (UniqueName: \"kubernetes.io/projected/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-kube-api-access-j79wm\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.578020 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b4a4d02-59a6-4f74-a4cc-2154d1115361-installation-pull-secrets\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.578634 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-audit\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.578653 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-encryption-config\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.578868 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-serving-cert\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.579232 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db53a48c-000d-4e92-929e-6c6bf561119a-trusted-ca\") pod \"console-operator-58897d9998-5xv2z\" (UID: \"db53a48c-000d-4e92-929e-6c6bf561119a\") " pod="openshift-console-operator/console-operator-58897d9998-5xv2z" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.579426 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.579847 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.580183 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-service-ca\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.580520 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4g45\" (UniqueName: \"kubernetes.io/projected/98112972-5121-4d93-8490-5156a1a2f3ca-kube-api-access-m4g45\") pod \"downloads-7954f5f757-2n4tf\" (UID: \"98112972-5121-4d93-8490-5156a1a2f3ca\") " pod="openshift-console/downloads-7954f5f757-2n4tf" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.580752 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46c9f2b1-15f1-418b-a086-5ea9972185a0-config\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.580830 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-xx5hs\" (UID: \"d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.580868 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-etcd-serving-ca\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.580903 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/cdc3c78e-177f-4fcd-ab8d-c333497848aa-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-nh2sn\" (UID: \"cdc3c78e-177f-4fcd-ab8d-c333497848aa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.581349 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fc590a1b-ba51-421c-a089-fb0def054fb2-metrics-tls\") pod \"dns-operator-744455d44c-pwdrv\" (UID: \"fc590a1b-ba51-421c-a089-fb0def054fb2\") " pod="openshift-dns-operator/dns-operator-744455d44c-pwdrv" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.581359 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e245069f-c15d-4ad4-9709-6bec4a52c8c8-config\") pod \"authentication-operator-69f744f599-jds5c\" (UID: \"e245069f-c15d-4ad4-9709-6bec4a52c8c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.581392 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46c9f2b1-15f1-418b-a086-5ea9972185a0-config\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.582305 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46c9f2b1-15f1-418b-a086-5ea9972185a0-serving-cert\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.582322 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-etcd-client\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.582372 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b78e04c-f222-4aae-be52-85abd4ecf391-config-volume\") pod \"dns-default-sbf5t\" (UID: \"9b78e04c-f222-4aae-be52-85abd4ecf391\") " pod="openshift-dns/dns-default-sbf5t" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.582411 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a357d8b-0e0e-4ae0-9365-4c4d96316ae2-serving-cert\") pod \"service-ca-operator-777779d784-4xl77\" (UID: \"1a357d8b-0e0e-4ae0-9365-4c4d96316ae2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xl77" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.582440 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d367fc6a-d4c4-4e63-81da-c547e07eea8d-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhq9n\" (UID: \"d367fc6a-d4c4-4e63-81da-c547e07eea8d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.582485 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-image-import-ca\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.582514 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/0f52f92d-a4bd-41b3-88dd-579a355f6e30-csi-data-dir\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.582565 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwp8x\" (UniqueName: \"kubernetes.io/projected/d518d443-b142-4235-8720-31d8f421c794-kube-api-access-dwp8x\") pod \"machine-config-server-m9qmb\" (UID: \"d518d443-b142-4235-8720-31d8f421c794\") " pod="openshift-machine-config-operator/machine-config-server-m9qmb" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.582595 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6401ae31-5292-4d4f-8d39-a5682c0d38ee-profile-collector-cert\") pod \"catalog-operator-68c6474976-tzdzl\" (UID: \"6401ae31-5292-4d4f-8d39-a5682c0d38ee\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.582627 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m95jf\" (UniqueName: \"kubernetes.io/projected/6401ae31-5292-4d4f-8d39-a5682c0d38ee-kube-api-access-m95jf\") pod \"catalog-operator-68c6474976-tzdzl\" (UID: \"6401ae31-5292-4d4f-8d39-a5682c0d38ee\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.582657 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db53a48c-000d-4e92-929e-6c6bf561119a-serving-cert\") pod \"console-operator-58897d9998-5xv2z\" (UID: \"db53a48c-000d-4e92-929e-6c6bf561119a\") " pod="openshift-console-operator/console-operator-58897d9998-5xv2z" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.582872 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e245069f-c15d-4ad4-9709-6bec4a52c8c8-config\") pod \"authentication-operator-69f744f599-jds5c\" (UID: \"e245069f-c15d-4ad4-9709-6bec4a52c8c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.582957 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04649429-b45f-41ff-8777-4e0b321664b9-serving-cert\") pod \"openshift-config-operator-7777fb866f-qnsbd\" (UID: \"04649429-b45f-41ff-8777-4e0b321664b9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.582976 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d518d443-b142-4235-8720-31d8f421c794-node-bootstrap-token\") pod \"machine-config-server-m9qmb\" (UID: \"d518d443-b142-4235-8720-31d8f421c794\") " pod="openshift-machine-config-operator/machine-config-server-m9qmb" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.583102 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/057f89ed-0870-4017-8a0c-a4808d3fb86b-webhook-cert\") pod \"packageserver-d55dfcdfc-8tt9s\" (UID: \"057f89ed-0870-4017-8a0c-a4808d3fb86b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.583193 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-etcd-client\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.583313 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-serving-cert\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.583371 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9541f10f-b16a-4dbe-8de2-fb4f381fdb94-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gnk2x\" (UID: \"9541f10f-b16a-4dbe-8de2-fb4f381fdb94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.583504 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-oauth-serving-cert\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.583952 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m272t\" (UniqueName: \"kubernetes.io/projected/3dafc1d6-a249-4d30-ad93-f065110076c8-kube-api-access-m272t\") pod \"control-plane-machine-set-operator-78cbb6b69f-ws5nw\" (UID: \"3dafc1d6-a249-4d30-ad93-f065110076c8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ws5nw" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.584013 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d518d443-b142-4235-8720-31d8f421c794-certs\") pod \"machine-config-server-m9qmb\" (UID: \"d518d443-b142-4235-8720-31d8f421c794\") " pod="openshift-machine-config-operator/machine-config-server-m9qmb" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.584119 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l92p\" (UniqueName: \"kubernetes.io/projected/fc590a1b-ba51-421c-a089-fb0def054fb2-kube-api-access-9l92p\") pod \"dns-operator-744455d44c-pwdrv\" (UID: \"fc590a1b-ba51-421c-a089-fb0def054fb2\") " pod="openshift-dns-operator/dns-operator-744455d44c-pwdrv" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.584187 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc099c2d-2785-4509-bd8b-29469341f56d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ndc8k\" (UID: \"bc099c2d-2785-4509-bd8b-29469341f56d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.584198 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-encryption-config\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.584275 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e245069f-c15d-4ad4-9709-6bec4a52c8c8-serving-cert\") pod \"authentication-operator-69f744f599-jds5c\" (UID: \"e245069f-c15d-4ad4-9709-6bec4a52c8c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.584304 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dfa49cf2-dc52-4c27-8486-ec1efec61744-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mjlld\" (UID: \"dfa49cf2-dc52-4c27-8486-ec1efec61744\") " pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.584477 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-oauth-serving-cert\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.584766 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b8fa45e-2136-44a1-9a50-f936b5c22417-serving-cert\") pod \"route-controller-manager-6576b87f9c-gmfbn\" (UID: \"2b8fa45e-2136-44a1-9a50-f936b5c22417\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.584931 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/46c9f2b1-15f1-418b-a086-5ea9972185a0-etcd-ca\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.584965 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8bc5b690-2e68-4d62-a0cf-349326b8e5e4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hl9wj\" (UID: \"8bc5b690-2e68-4d62-a0cf-349326b8e5e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.585084 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scc9t\" (UniqueName: \"kubernetes.io/projected/0f52f92d-a4bd-41b3-88dd-579a355f6e30-kube-api-access-scc9t\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.585116 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97d5e809-690b-44bc-9df0-1af46832e149-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lmzk2\" (UID: \"97d5e809-690b-44bc-9df0-1af46832e149\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.585145 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6lp4\" (UniqueName: \"kubernetes.io/projected/3e0b793d-24a2-4c73-9357-fe193c32f71b-kube-api-access-q6lp4\") pod \"machine-config-operator-74547568cd-w4f4k\" (UID: \"3e0b793d-24a2-4c73-9357-fe193c32f71b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.585163 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/008442d1-8ae3-4669-945d-d7f3389855a2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-t6c54\" (UID: \"008442d1-8ae3-4669-945d-d7f3389855a2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.585913 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-registry-tls\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.586031 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db53a48c-000d-4e92-929e-6c6bf561119a-serving-cert\") pod \"console-operator-58897d9998-5xv2z\" (UID: \"db53a48c-000d-4e92-929e-6c6bf561119a\") " pod="openshift-console-operator/console-operator-58897d9998-5xv2z" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.586333 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-image-import-ca\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.586537 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/46c9f2b1-15f1-418b-a086-5ea9972185a0-etcd-ca\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.586782 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-oauth-config\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.588471 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-serving-cert\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.588951 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc099c2d-2785-4509-bd8b-29469341f56d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ndc8k\" (UID: \"bc099c2d-2785-4509-bd8b-29469341f56d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.589584 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-serving-cert\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.590371 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-etcd-client\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.591514 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-xx5hs\" (UID: \"d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.592598 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e245069f-c15d-4ad4-9709-6bec4a52c8c8-serving-cert\") pod \"authentication-operator-69f744f599-jds5c\" (UID: \"e245069f-c15d-4ad4-9709-6bec4a52c8c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.614128 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" event={"ID":"8b44d028-0edc-4477-9f01-236c6d59d9e8","Type":"ContainerStarted","Data":"9ef51e377b2fff9d194cc8678447e9b1a55ecbbf4d21123b11551f49759c5ea3"} Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.614179 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" event={"ID":"8b44d028-0edc-4477-9f01-236c6d59d9e8","Type":"ContainerStarted","Data":"a049ba81939ac38bdd431ca7d531681efc12eab1e258d8cc4d4dc2072abb0dfc"} Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.615112 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmknd\" (UniqueName: \"kubernetes.io/projected/f487b7f4-c24b-431e-8868-a0285c8ed7bc-kube-api-access-qmknd\") pod \"console-f9d7485db-7x2qp\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.629898 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4rvqh"] Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.635796 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv48m\" (UniqueName: \"kubernetes.io/projected/04649429-b45f-41ff-8777-4e0b321664b9-kube-api-access-wv48m\") pod \"openshift-config-operator-7777fb866f-qnsbd\" (UID: \"04649429-b45f-41ff-8777-4e0b321664b9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.652349 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c9r4\" (UniqueName: \"kubernetes.io/projected/2b8fa45e-2136-44a1-9a50-f936b5c22417-kube-api-access-4c9r4\") pod \"route-controller-manager-6576b87f9c-gmfbn\" (UID: \"2b8fa45e-2136-44a1-9a50-f936b5c22417\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.679153 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.679590 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cdc3c78e-177f-4fcd-ab8d-c333497848aa-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-nh2sn\" (UID: \"cdc3c78e-177f-4fcd-ab8d-c333497848aa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.687834 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:51 crc kubenswrapper[4814]: E0227 16:26:51.688049 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:52.188006924 +0000 UTC m=+224.640631754 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688119 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx4gm\" (UniqueName: \"kubernetes.io/projected/3b269e43-ebd6-4f71-8963-669c60763efa-kube-api-access-zx4gm\") pod \"machine-config-controller-84d6567774-g4vb2\" (UID: \"3b269e43-ebd6-4f71-8963-669c60763efa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688183 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/13240f27-5cc2-4e90-9992-24f6beca4212-default-certificate\") pod \"router-default-5444994796-wv7m2\" (UID: \"13240f27-5cc2-4e90-9992-24f6beca4212\") " pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688235 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6401ae31-5292-4d4f-8d39-a5682c0d38ee-srv-cert\") pod \"catalog-operator-68c6474976-tzdzl\" (UID: \"6401ae31-5292-4d4f-8d39-a5682c0d38ee\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688314 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55b73477-5f70-4dcc-a9aa-3e22ac4efb54-cert\") pod \"ingress-canary-6hpl2\" (UID: \"55b73477-5f70-4dcc-a9aa-3e22ac4efb54\") " pod="openshift-ingress-canary/ingress-canary-6hpl2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688374 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13240f27-5cc2-4e90-9992-24f6beca4212-metrics-certs\") pod \"router-default-5444994796-wv7m2\" (UID: \"13240f27-5cc2-4e90-9992-24f6beca4212\") " pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688397 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggb4k\" (UniqueName: \"kubernetes.io/projected/d87f2169-8a70-42de-aaf5-7728ff95fa50-kube-api-access-ggb4k\") pod \"collect-profiles-29536815-pqwsb\" (UID: \"d87f2169-8a70-42de-aaf5-7728ff95fa50\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688426 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3b269e43-ebd6-4f71-8963-669c60763efa-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-g4vb2\" (UID: \"3b269e43-ebd6-4f71-8963-669c60763efa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688460 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/057f89ed-0870-4017-8a0c-a4808d3fb86b-tmpfs\") pod \"packageserver-d55dfcdfc-8tt9s\" (UID: \"057f89ed-0870-4017-8a0c-a4808d3fb86b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688481 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13240f27-5cc2-4e90-9992-24f6beca4212-service-ca-bundle\") pod \"router-default-5444994796-wv7m2\" (UID: \"13240f27-5cc2-4e90-9992-24f6beca4212\") " pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688541 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nstsm\" (UniqueName: \"kubernetes.io/projected/634e6fbd-87f9-4205-b01d-90be6eb4a015-kube-api-access-nstsm\") pod \"migrator-59844c95c7-twcnw\" (UID: \"634e6fbd-87f9-4205-b01d-90be6eb4a015\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twcnw" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688573 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ba54dbcb-efbd-467b-aa4e-994fde5e1c14-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ncx8v\" (UID: \"ba54dbcb-efbd-467b-aa4e-994fde5e1c14\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ncx8v" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688596 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/0f52f92d-a4bd-41b3-88dd-579a355f6e30-mountpoint-dir\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688618 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9541f10f-b16a-4dbe-8de2-fb4f381fdb94-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gnk2x\" (UID: \"9541f10f-b16a-4dbe-8de2-fb4f381fdb94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688639 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w6zl\" (UniqueName: \"kubernetes.io/projected/95913cb4-39f1-44c0-ac49-0a2d51047679-kube-api-access-8w6zl\") pod \"auto-csr-approver-29536826-dgwn4\" (UID: \"95913cb4-39f1-44c0-ac49-0a2d51047679\") " pod="openshift-infra/auto-csr-approver-29536826-dgwn4" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688665 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48wrk\" (UniqueName: \"kubernetes.io/projected/8676c139-d437-4f7a-9234-de7c8a1b8dd1-kube-api-access-48wrk\") pod \"service-ca-9c57cc56f-qzwns\" (UID: \"8676c139-d437-4f7a-9234-de7c8a1b8dd1\") " pod="openshift-service-ca/service-ca-9c57cc56f-qzwns" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688690 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d367fc6a-d4c4-4e63-81da-c547e07eea8d-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhq9n\" (UID: \"d367fc6a-d4c4-4e63-81da-c547e07eea8d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688709 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d87f2169-8a70-42de-aaf5-7728ff95fa50-config-volume\") pod \"collect-profiles-29536815-pqwsb\" (UID: \"d87f2169-8a70-42de-aaf5-7728ff95fa50\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688726 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/13240f27-5cc2-4e90-9992-24f6beca4212-stats-auth\") pod \"router-default-5444994796-wv7m2\" (UID: \"13240f27-5cc2-4e90-9992-24f6beca4212\") " pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688758 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d87f2169-8a70-42de-aaf5-7728ff95fa50-secret-volume\") pod \"collect-profiles-29536815-pqwsb\" (UID: \"d87f2169-8a70-42de-aaf5-7728ff95fa50\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688801 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfa49cf2-dc52-4c27-8486-ec1efec61744-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mjlld\" (UID: \"dfa49cf2-dc52-4c27-8486-ec1efec61744\") " pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688830 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr6lk\" (UniqueName: \"kubernetes.io/projected/13240f27-5cc2-4e90-9992-24f6beca4212-kube-api-access-lr6lk\") pod \"router-default-5444994796-wv7m2\" (UID: \"13240f27-5cc2-4e90-9992-24f6beca4212\") " pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688917 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d367fc6a-d4c4-4e63-81da-c547e07eea8d-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhq9n\" (UID: \"d367fc6a-d4c4-4e63-81da-c547e07eea8d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688950 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/0f52f92d-a4bd-41b3-88dd-579a355f6e30-socket-dir\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.688983 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3dafc1d6-a249-4d30-ad93-f065110076c8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ws5nw\" (UID: \"3dafc1d6-a249-4d30-ad93-f065110076c8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ws5nw" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689012 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8676c139-d437-4f7a-9234-de7c8a1b8dd1-signing-key\") pod \"service-ca-9c57cc56f-qzwns\" (UID: \"8676c139-d437-4f7a-9234-de7c8a1b8dd1\") " pod="openshift-service-ca/service-ca-9c57cc56f-qzwns" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689043 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld5b5\" (UniqueName: \"kubernetes.io/projected/55b73477-5f70-4dcc-a9aa-3e22ac4efb54-kube-api-access-ld5b5\") pod \"ingress-canary-6hpl2\" (UID: \"55b73477-5f70-4dcc-a9aa-3e22ac4efb54\") " pod="openshift-ingress-canary/ingress-canary-6hpl2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689073 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgcgk\" (UniqueName: \"kubernetes.io/projected/ba54dbcb-efbd-467b-aa4e-994fde5e1c14-kube-api-access-wgcgk\") pod \"multus-admission-controller-857f4d67dd-ncx8v\" (UID: \"ba54dbcb-efbd-467b-aa4e-994fde5e1c14\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ncx8v" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689109 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8676c139-d437-4f7a-9234-de7c8a1b8dd1-signing-cabundle\") pod \"service-ca-9c57cc56f-qzwns\" (UID: \"8676c139-d437-4f7a-9234-de7c8a1b8dd1\") " pod="openshift-service-ca/service-ca-9c57cc56f-qzwns" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689155 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7kwh\" (UniqueName: \"kubernetes.io/projected/1a357d8b-0e0e-4ae0-9365-4c4d96316ae2-kube-api-access-p7kwh\") pod \"service-ca-operator-777779d784-4xl77\" (UID: \"1a357d8b-0e0e-4ae0-9365-4c4d96316ae2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xl77" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689203 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97d5e809-690b-44bc-9df0-1af46832e149-config\") pod \"kube-controller-manager-operator-78b949d7b-lmzk2\" (UID: \"97d5e809-690b-44bc-9df0-1af46832e149\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689236 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d32d7bef-558f-4dfd-9260-408cb47fc461-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rs9mx\" (UID: \"d32d7bef-558f-4dfd-9260-408cb47fc461\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689286 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/0f52f92d-a4bd-41b3-88dd-579a355f6e30-registration-dir\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689324 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3e0b793d-24a2-4c73-9357-fe193c32f71b-proxy-tls\") pod \"machine-config-operator-74547568cd-w4f4k\" (UID: \"3e0b793d-24a2-4c73-9357-fe193c32f71b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689369 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32d7bef-558f-4dfd-9260-408cb47fc461-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rs9mx\" (UID: \"d32d7bef-558f-4dfd-9260-408cb47fc461\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689406 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsmlh\" (UniqueName: \"kubernetes.io/projected/057f89ed-0870-4017-8a0c-a4808d3fb86b-kube-api-access-qsmlh\") pod \"packageserver-d55dfcdfc-8tt9s\" (UID: \"057f89ed-0870-4017-8a0c-a4808d3fb86b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689444 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a357d8b-0e0e-4ae0-9365-4c4d96316ae2-config\") pod \"service-ca-operator-777779d784-4xl77\" (UID: \"1a357d8b-0e0e-4ae0-9365-4c4d96316ae2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xl77" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689470 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8bc5b690-2e68-4d62-a0cf-349326b8e5e4-srv-cert\") pod \"olm-operator-6b444d44fb-hl9wj\" (UID: \"8bc5b690-2e68-4d62-a0cf-349326b8e5e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689515 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmsmw\" (UniqueName: \"kubernetes.io/projected/dfa49cf2-dc52-4c27-8486-ec1efec61744-kube-api-access-qmsmw\") pod \"marketplace-operator-79b997595-mjlld\" (UID: \"dfa49cf2-dc52-4c27-8486-ec1efec61744\") " pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689555 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt2g7\" (UniqueName: \"kubernetes.io/projected/8bc5b690-2e68-4d62-a0cf-349326b8e5e4-kube-api-access-qt2g7\") pod \"olm-operator-6b444d44fb-hl9wj\" (UID: \"8bc5b690-2e68-4d62-a0cf-349326b8e5e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689633 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b78e04c-f222-4aae-be52-85abd4ecf391-config-volume\") pod \"dns-default-sbf5t\" (UID: \"9b78e04c-f222-4aae-be52-85abd4ecf391\") " pod="openshift-dns/dns-default-sbf5t" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689641 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3b269e43-ebd6-4f71-8963-669c60763efa-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-g4vb2\" (UID: \"3b269e43-ebd6-4f71-8963-669c60763efa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689663 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a357d8b-0e0e-4ae0-9365-4c4d96316ae2-serving-cert\") pod \"service-ca-operator-777779d784-4xl77\" (UID: \"1a357d8b-0e0e-4ae0-9365-4c4d96316ae2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xl77" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689745 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d367fc6a-d4c4-4e63-81da-c547e07eea8d-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhq9n\" (UID: \"d367fc6a-d4c4-4e63-81da-c547e07eea8d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689771 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m95jf\" (UniqueName: \"kubernetes.io/projected/6401ae31-5292-4d4f-8d39-a5682c0d38ee-kube-api-access-m95jf\") pod \"catalog-operator-68c6474976-tzdzl\" (UID: \"6401ae31-5292-4d4f-8d39-a5682c0d38ee\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689800 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/0f52f92d-a4bd-41b3-88dd-579a355f6e30-csi-data-dir\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689821 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwp8x\" (UniqueName: \"kubernetes.io/projected/d518d443-b142-4235-8720-31d8f421c794-kube-api-access-dwp8x\") pod \"machine-config-server-m9qmb\" (UID: \"d518d443-b142-4235-8720-31d8f421c794\") " pod="openshift-machine-config-operator/machine-config-server-m9qmb" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689838 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6401ae31-5292-4d4f-8d39-a5682c0d38ee-profile-collector-cert\") pod \"catalog-operator-68c6474976-tzdzl\" (UID: \"6401ae31-5292-4d4f-8d39-a5682c0d38ee\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689866 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d518d443-b142-4235-8720-31d8f421c794-node-bootstrap-token\") pod \"machine-config-server-m9qmb\" (UID: \"d518d443-b142-4235-8720-31d8f421c794\") " pod="openshift-machine-config-operator/machine-config-server-m9qmb" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689887 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/057f89ed-0870-4017-8a0c-a4808d3fb86b-webhook-cert\") pod \"packageserver-d55dfcdfc-8tt9s\" (UID: \"057f89ed-0870-4017-8a0c-a4808d3fb86b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689911 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9541f10f-b16a-4dbe-8de2-fb4f381fdb94-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gnk2x\" (UID: \"9541f10f-b16a-4dbe-8de2-fb4f381fdb94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689932 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m272t\" (UniqueName: \"kubernetes.io/projected/3dafc1d6-a249-4d30-ad93-f065110076c8-kube-api-access-m272t\") pod \"control-plane-machine-set-operator-78cbb6b69f-ws5nw\" (UID: \"3dafc1d6-a249-4d30-ad93-f065110076c8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ws5nw" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689950 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d518d443-b142-4235-8720-31d8f421c794-certs\") pod \"machine-config-server-m9qmb\" (UID: \"d518d443-b142-4235-8720-31d8f421c794\") " pod="openshift-machine-config-operator/machine-config-server-m9qmb" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.689999 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dfa49cf2-dc52-4c27-8486-ec1efec61744-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mjlld\" (UID: \"dfa49cf2-dc52-4c27-8486-ec1efec61744\") " pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690021 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8bc5b690-2e68-4d62-a0cf-349326b8e5e4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hl9wj\" (UID: \"8bc5b690-2e68-4d62-a0cf-349326b8e5e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690047 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scc9t\" (UniqueName: \"kubernetes.io/projected/0f52f92d-a4bd-41b3-88dd-579a355f6e30-kube-api-access-scc9t\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690069 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97d5e809-690b-44bc-9df0-1af46832e149-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lmzk2\" (UID: \"97d5e809-690b-44bc-9df0-1af46832e149\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690089 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6lp4\" (UniqueName: \"kubernetes.io/projected/3e0b793d-24a2-4c73-9357-fe193c32f71b-kube-api-access-q6lp4\") pod \"machine-config-operator-74547568cd-w4f4k\" (UID: \"3e0b793d-24a2-4c73-9357-fe193c32f71b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690115 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7dc9\" (UniqueName: \"kubernetes.io/projected/5768bed9-9187-4bef-ae5e-bc003793d741-kube-api-access-t7dc9\") pod \"package-server-manager-789f6589d5-6c5hj\" (UID: \"5768bed9-9187-4bef-ae5e-bc003793d741\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690135 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3b269e43-ebd6-4f71-8963-669c60763efa-proxy-tls\") pod \"machine-config-controller-84d6567774-g4vb2\" (UID: \"3b269e43-ebd6-4f71-8963-669c60763efa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690154 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/0f52f92d-a4bd-41b3-88dd-579a355f6e30-plugins-dir\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690172 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3e0b793d-24a2-4c73-9357-fe193c32f71b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-w4f4k\" (UID: \"3e0b793d-24a2-4c73-9357-fe193c32f71b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690192 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9b78e04c-f222-4aae-be52-85abd4ecf391-metrics-tls\") pod \"dns-default-sbf5t\" (UID: \"9b78e04c-f222-4aae-be52-85abd4ecf391\") " pod="openshift-dns/dns-default-sbf5t" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690213 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97d5e809-690b-44bc-9df0-1af46832e149-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lmzk2\" (UID: \"97d5e809-690b-44bc-9df0-1af46832e149\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690228 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/057f89ed-0870-4017-8a0c-a4808d3fb86b-apiservice-cert\") pod \"packageserver-d55dfcdfc-8tt9s\" (UID: \"057f89ed-0870-4017-8a0c-a4808d3fb86b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690248 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3e0b793d-24a2-4c73-9357-fe193c32f71b-images\") pod \"machine-config-operator-74547568cd-w4f4k\" (UID: \"3e0b793d-24a2-4c73-9357-fe193c32f71b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690287 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690313 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5768bed9-9187-4bef-ae5e-bc003793d741-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-6c5hj\" (UID: \"5768bed9-9187-4bef-ae5e-bc003793d741\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690335 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpbj9\" (UniqueName: \"kubernetes.io/projected/9541f10f-b16a-4dbe-8de2-fb4f381fdb94-kube-api-access-gpbj9\") pod \"kube-storage-version-migrator-operator-b67b599dd-gnk2x\" (UID: \"9541f10f-b16a-4dbe-8de2-fb4f381fdb94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690352 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32d7bef-558f-4dfd-9260-408cb47fc461-config\") pod \"kube-apiserver-operator-766d6c64bb-rs9mx\" (UID: \"d32d7bef-558f-4dfd-9260-408cb47fc461\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690368 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbhfj\" (UniqueName: \"kubernetes.io/projected/9b78e04c-f222-4aae-be52-85abd4ecf391-kube-api-access-cbhfj\") pod \"dns-default-sbf5t\" (UID: \"9b78e04c-f222-4aae-be52-85abd4ecf391\") " pod="openshift-dns/dns-default-sbf5t" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.690633 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/0f52f92d-a4bd-41b3-88dd-579a355f6e30-csi-data-dir\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.691935 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d87f2169-8a70-42de-aaf5-7728ff95fa50-config-volume\") pod \"collect-profiles-29536815-pqwsb\" (UID: \"d87f2169-8a70-42de-aaf5-7728ff95fa50\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.692958 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/057f89ed-0870-4017-8a0c-a4808d3fb86b-tmpfs\") pod \"packageserver-d55dfcdfc-8tt9s\" (UID: \"057f89ed-0870-4017-8a0c-a4808d3fb86b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.693002 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97d5e809-690b-44bc-9df0-1af46832e149-config\") pod \"kube-controller-manager-operator-78b949d7b-lmzk2\" (UID: \"97d5e809-690b-44bc-9df0-1af46832e149\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.694658 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13240f27-5cc2-4e90-9992-24f6beca4212-service-ca-bundle\") pod \"router-default-5444994796-wv7m2\" (UID: \"13240f27-5cc2-4e90-9992-24f6beca4212\") " pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.695403 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/0f52f92d-a4bd-41b3-88dd-579a355f6e30-registration-dir\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.695859 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d367fc6a-d4c4-4e63-81da-c547e07eea8d-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhq9n\" (UID: \"d367fc6a-d4c4-4e63-81da-c547e07eea8d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.698384 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/0f52f92d-a4bd-41b3-88dd-579a355f6e30-plugins-dir\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.698725 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3e0b793d-24a2-4c73-9357-fe193c32f71b-images\") pod \"machine-config-operator-74547568cd-w4f4k\" (UID: \"3e0b793d-24a2-4c73-9357-fe193c32f71b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" Feb 27 16:26:51 crc kubenswrapper[4814]: E0227 16:26:51.699119 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:52.199099657 +0000 UTC m=+224.651724487 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.699224 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3e0b793d-24a2-4c73-9357-fe193c32f71b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-w4f4k\" (UID: \"3e0b793d-24a2-4c73-9357-fe193c32f71b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.700167 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a357d8b-0e0e-4ae0-9365-4c4d96316ae2-config\") pod \"service-ca-operator-777779d784-4xl77\" (UID: \"1a357d8b-0e0e-4ae0-9365-4c4d96316ae2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xl77" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.700507 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9541f10f-b16a-4dbe-8de2-fb4f381fdb94-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gnk2x\" (UID: \"9541f10f-b16a-4dbe-8de2-fb4f381fdb94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.701442 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8676c139-d437-4f7a-9234-de7c8a1b8dd1-signing-cabundle\") pod \"service-ca-9c57cc56f-qzwns\" (UID: \"8676c139-d437-4f7a-9234-de7c8a1b8dd1\") " pod="openshift-service-ca/service-ca-9c57cc56f-qzwns" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.701612 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a357d8b-0e0e-4ae0-9365-4c4d96316ae2-serving-cert\") pod \"service-ca-operator-777779d784-4xl77\" (UID: \"1a357d8b-0e0e-4ae0-9365-4c4d96316ae2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xl77" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.701756 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/0f52f92d-a4bd-41b3-88dd-579a355f6e30-mountpoint-dir\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.702544 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/0f52f92d-a4bd-41b3-88dd-579a355f6e30-socket-dir\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.703133 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfa49cf2-dc52-4c27-8486-ec1efec61744-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mjlld\" (UID: \"dfa49cf2-dc52-4c27-8486-ec1efec61744\") " pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.703389 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32d7bef-558f-4dfd-9260-408cb47fc461-config\") pod \"kube-apiserver-operator-766d6c64bb-rs9mx\" (UID: \"d32d7bef-558f-4dfd-9260-408cb47fc461\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.704194 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55b73477-5f70-4dcc-a9aa-3e22ac4efb54-cert\") pod \"ingress-canary-6hpl2\" (UID: \"55b73477-5f70-4dcc-a9aa-3e22ac4efb54\") " pod="openshift-ingress-canary/ingress-canary-6hpl2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.704208 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6401ae31-5292-4d4f-8d39-a5682c0d38ee-srv-cert\") pod \"catalog-operator-68c6474976-tzdzl\" (UID: \"6401ae31-5292-4d4f-8d39-a5682c0d38ee\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.705571 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b78e04c-f222-4aae-be52-85abd4ecf391-config-volume\") pod \"dns-default-sbf5t\" (UID: \"9b78e04c-f222-4aae-be52-85abd4ecf391\") " pod="openshift-dns/dns-default-sbf5t" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.706199 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/13240f27-5cc2-4e90-9992-24f6beca4212-default-certificate\") pod \"router-default-5444994796-wv7m2\" (UID: \"13240f27-5cc2-4e90-9992-24f6beca4212\") " pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.706281 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d367fc6a-d4c4-4e63-81da-c547e07eea8d-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhq9n\" (UID: \"d367fc6a-d4c4-4e63-81da-c547e07eea8d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.706710 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97d5e809-690b-44bc-9df0-1af46832e149-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lmzk2\" (UID: \"97d5e809-690b-44bc-9df0-1af46832e149\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.706809 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlsxd\" (UniqueName: \"kubernetes.io/projected/46c9f2b1-15f1-418b-a086-5ea9972185a0-kube-api-access-mlsxd\") pod \"etcd-operator-b45778765-9vmj5\" (UID: \"46c9f2b1-15f1-418b-a086-5ea9972185a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.708064 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8bc5b690-2e68-4d62-a0cf-349326b8e5e4-srv-cert\") pod \"olm-operator-6b444d44fb-hl9wj\" (UID: \"8bc5b690-2e68-4d62-a0cf-349326b8e5e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.709225 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32d7bef-558f-4dfd-9260-408cb47fc461-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rs9mx\" (UID: \"d32d7bef-558f-4dfd-9260-408cb47fc461\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.710085 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3dafc1d6-a249-4d30-ad93-f065110076c8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ws5nw\" (UID: \"3dafc1d6-a249-4d30-ad93-f065110076c8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ws5nw" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.710295 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3e0b793d-24a2-4c73-9357-fe193c32f71b-proxy-tls\") pod \"machine-config-operator-74547568cd-w4f4k\" (UID: \"3e0b793d-24a2-4c73-9357-fe193c32f71b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.710700 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3b269e43-ebd6-4f71-8963-669c60763efa-proxy-tls\") pod \"machine-config-controller-84d6567774-g4vb2\" (UID: \"3b269e43-ebd6-4f71-8963-669c60763efa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.710781 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8676c139-d437-4f7a-9234-de7c8a1b8dd1-signing-key\") pod \"service-ca-9c57cc56f-qzwns\" (UID: \"8676c139-d437-4f7a-9234-de7c8a1b8dd1\") " pod="openshift-service-ca/service-ca-9c57cc56f-qzwns" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.711096 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13240f27-5cc2-4e90-9992-24f6beca4212-metrics-certs\") pod \"router-default-5444994796-wv7m2\" (UID: \"13240f27-5cc2-4e90-9992-24f6beca4212\") " pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.711145 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/13240f27-5cc2-4e90-9992-24f6beca4212-stats-auth\") pod \"router-default-5444994796-wv7m2\" (UID: \"13240f27-5cc2-4e90-9992-24f6beca4212\") " pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.711322 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/057f89ed-0870-4017-8a0c-a4808d3fb86b-apiservice-cert\") pod \"packageserver-d55dfcdfc-8tt9s\" (UID: \"057f89ed-0870-4017-8a0c-a4808d3fb86b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.711306 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d87f2169-8a70-42de-aaf5-7728ff95fa50-secret-volume\") pod \"collect-profiles-29536815-pqwsb\" (UID: \"d87f2169-8a70-42de-aaf5-7728ff95fa50\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.711531 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dfa49cf2-dc52-4c27-8486-ec1efec61744-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mjlld\" (UID: \"dfa49cf2-dc52-4c27-8486-ec1efec61744\") " pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.712181 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6401ae31-5292-4d4f-8d39-a5682c0d38ee-profile-collector-cert\") pod \"catalog-operator-68c6474976-tzdzl\" (UID: \"6401ae31-5292-4d4f-8d39-a5682c0d38ee\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.713757 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5768bed9-9187-4bef-ae5e-bc003793d741-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-6c5hj\" (UID: \"5768bed9-9187-4bef-ae5e-bc003793d741\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.715115 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.716017 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9b78e04c-f222-4aae-be52-85abd4ecf391-metrics-tls\") pod \"dns-default-sbf5t\" (UID: \"9b78e04c-f222-4aae-be52-85abd4ecf391\") " pod="openshift-dns/dns-default-sbf5t" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.716080 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9541f10f-b16a-4dbe-8de2-fb4f381fdb94-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gnk2x\" (UID: \"9541f10f-b16a-4dbe-8de2-fb4f381fdb94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.717829 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d518d443-b142-4235-8720-31d8f421c794-node-bootstrap-token\") pod \"machine-config-server-m9qmb\" (UID: \"d518d443-b142-4235-8720-31d8f421c794\") " pod="openshift-machine-config-operator/machine-config-server-m9qmb" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.723068 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8bc5b690-2e68-4d62-a0cf-349326b8e5e4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hl9wj\" (UID: \"8bc5b690-2e68-4d62-a0cf-349326b8e5e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.724514 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq2xf\" (UniqueName: \"kubernetes.io/projected/bc099c2d-2785-4509-bd8b-29469341f56d-kube-api-access-cq2xf\") pod \"machine-api-operator-5694c8668f-ndc8k\" (UID: \"bc099c2d-2785-4509-bd8b-29469341f56d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.724551 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n"] Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.725222 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-k4p69"] Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.725924 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ba54dbcb-efbd-467b-aa4e-994fde5e1c14-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ncx8v\" (UID: \"ba54dbcb-efbd-467b-aa4e-994fde5e1c14\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ncx8v" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.726319 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4m8vp"] Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.727134 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/057f89ed-0870-4017-8a0c-a4808d3fb86b-webhook-cert\") pod \"packageserver-d55dfcdfc-8tt9s\" (UID: \"057f89ed-0870-4017-8a0c-a4808d3fb86b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.727386 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d518d443-b142-4235-8720-31d8f421c794-certs\") pod \"machine-config-server-m9qmb\" (UID: \"d518d443-b142-4235-8720-31d8f421c794\") " pod="openshift-machine-config-operator/machine-config-server-m9qmb" Feb 27 16:26:51 crc kubenswrapper[4814]: W0227 16:26:51.733139 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29babfee_4ec4_4828_90c6_caf8fb0647b2.slice/crio-cf00bfbcfdff6c29605896888c870bd2c4c376a1036dc8186cd1c3d9b23f1aa0 WatchSource:0}: Error finding container cf00bfbcfdff6c29605896888c870bd2c4c376a1036dc8186cd1c3d9b23f1aa0: Status 404 returned error can't find the container with id cf00bfbcfdff6c29605896888c870bd2c4c376a1036dc8186cd1c3d9b23f1aa0 Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.733858 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djkd5\" (UniqueName: \"kubernetes.io/projected/e245069f-c15d-4ad4-9709-6bec4a52c8c8-kube-api-access-djkd5\") pod \"authentication-operator-69f744f599-jds5c\" (UID: \"e245069f-c15d-4ad4-9709-6bec4a52c8c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.738539 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.777979 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8tlp\" (UniqueName: \"kubernetes.io/projected/d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c-kube-api-access-t8tlp\") pod \"openshift-controller-manager-operator-756b6f6bc6-xx5hs\" (UID: \"d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.779130 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfmj9\" (UniqueName: \"kubernetes.io/projected/008442d1-8ae3-4669-945d-d7f3389855a2-kube-api-access-lfmj9\") pod \"openshift-apiserver-operator-796bbdcf4f-t6c54\" (UID: \"008442d1-8ae3-4669-945d-d7f3389855a2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.779424 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.791636 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:51 crc kubenswrapper[4814]: E0227 16:26:51.791781 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:52.291754517 +0000 UTC m=+224.744379347 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.792319 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: E0227 16:26:51.792741 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:52.29271975 +0000 UTC m=+224.745344660 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.800381 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7cqz\" (UniqueName: \"kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-kube-api-access-m7cqz\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.813451 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-bound-sa-token\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.824033 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.835743 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8dbk\" (UniqueName: \"kubernetes.io/projected/2dc05d09-1fba-4a2e-93fe-3db86e8ff333-kube-api-access-d8dbk\") pod \"apiserver-7bbb656c7d-46nl6\" (UID: \"2dc05d09-1fba-4a2e-93fe-3db86e8ff333\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.836468 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.858135 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j79wm\" (UniqueName: \"kubernetes.io/projected/e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b-kube-api-access-j79wm\") pod \"apiserver-76f77b778f-pbsmc\" (UID: \"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b\") " pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.865623 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.876054 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4g45\" (UniqueName: \"kubernetes.io/projected/98112972-5121-4d93-8490-5156a1a2f3ca-kube-api-access-m4g45\") pod \"downloads-7954f5f757-2n4tf\" (UID: \"98112972-5121-4d93-8490-5156a1a2f3ca\") " pod="openshift-console/downloads-7954f5f757-2n4tf" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.881871 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn"] Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.896806 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:51 crc kubenswrapper[4814]: E0227 16:26:51.897320 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:52.397305401 +0000 UTC m=+224.849930231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.901646 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph5x9\" (UniqueName: \"kubernetes.io/projected/db53a48c-000d-4e92-929e-6c6bf561119a-kube-api-access-ph5x9\") pod \"console-operator-58897d9998-5xv2z\" (UID: \"db53a48c-000d-4e92-929e-6c6bf561119a\") " pod="openshift-console-operator/console-operator-58897d9998-5xv2z" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.917555 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.930245 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l92p\" (UniqueName: \"kubernetes.io/projected/fc590a1b-ba51-421c-a089-fb0def054fb2-kube-api-access-9l92p\") pod \"dns-operator-744455d44c-pwdrv\" (UID: \"fc590a1b-ba51-421c-a089-fb0def054fb2\") " pod="openshift-dns-operator/dns-operator-744455d44c-pwdrv" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.936770 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9vmj5"] Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.956669 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54klp\" (UniqueName: \"kubernetes.io/projected/cdc3c78e-177f-4fcd-ab8d-c333497848aa-kube-api-access-54klp\") pod \"cluster-image-registry-operator-dc59b4c8b-nh2sn\" (UID: \"cdc3c78e-177f-4fcd-ab8d-c333497848aa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.974336 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w6zl\" (UniqueName: \"kubernetes.io/projected/95913cb4-39f1-44c0-ac49-0a2d51047679-kube-api-access-8w6zl\") pod \"auto-csr-approver-29536826-dgwn4\" (UID: \"95913cb4-39f1-44c0-ac49-0a2d51047679\") " pod="openshift-infra/auto-csr-approver-29536826-dgwn4" Feb 27 16:26:51 crc kubenswrapper[4814]: I0227 16:26:51.985631 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7x2qp"] Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.006076 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:26:52 crc kubenswrapper[4814]: E0227 16:26:52.006740 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:52.506718919 +0000 UTC m=+224.959343739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.006122 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.006897 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536826-dgwn4" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.006224 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48wrk\" (UniqueName: \"kubernetes.io/projected/8676c139-d437-4f7a-9234-de7c8a1b8dd1-kube-api-access-48wrk\") pod \"service-ca-9c57cc56f-qzwns\" (UID: \"8676c139-d437-4f7a-9234-de7c8a1b8dd1\") " pod="openshift-service-ca/service-ca-9c57cc56f-qzwns" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.020739 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbhfj\" (UniqueName: \"kubernetes.io/projected/9b78e04c-f222-4aae-be52-85abd4ecf391-kube-api-access-cbhfj\") pod \"dns-default-sbf5t\" (UID: \"9b78e04c-f222-4aae-be52-85abd4ecf391\") " pod="openshift-dns/dns-default-sbf5t" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.022478 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.030885 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-qzwns" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.038982 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d367fc6a-d4c4-4e63-81da-c547e07eea8d-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhq9n\" (UID: \"d367fc6a-d4c4-4e63-81da-c547e07eea8d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.054722 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-sbf5t" Feb 27 16:26:52 crc kubenswrapper[4814]: W0227 16:26:52.059792 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf487b7f4_c24b_431e_8868_a0285c8ed7bc.slice/crio-e461de5943467cdbf6f7248a16c5dfd1c11ff1400fc8140953caf4be6f92c322 WatchSource:0}: Error finding container e461de5943467cdbf6f7248a16c5dfd1c11ff1400fc8140953caf4be6f92c322: Status 404 returned error can't find the container with id e461de5943467cdbf6f7248a16c5dfd1c11ff1400fc8140953caf4be6f92c322 Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.063059 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m95jf\" (UniqueName: \"kubernetes.io/projected/6401ae31-5292-4d4f-8d39-a5682c0d38ee-kube-api-access-m95jf\") pod \"catalog-operator-68c6474976-tzdzl\" (UID: \"6401ae31-5292-4d4f-8d39-a5682c0d38ee\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.078183 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwp8x\" (UniqueName: \"kubernetes.io/projected/d518d443-b142-4235-8720-31d8f421c794-kube-api-access-dwp8x\") pod \"machine-config-server-m9qmb\" (UID: \"d518d443-b142-4235-8720-31d8f421c794\") " pod="openshift-machine-config-operator/machine-config-server-m9qmb" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.091462 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-pwdrv" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.096744 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7dc9\" (UniqueName: \"kubernetes.io/projected/5768bed9-9187-4bef-ae5e-bc003793d741-kube-api-access-t7dc9\") pod \"package-server-manager-789f6589d5-6c5hj\" (UID: \"5768bed9-9187-4bef-ae5e-bc003793d741\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.108118 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:52 crc kubenswrapper[4814]: E0227 16:26:52.108581 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:52.608551386 +0000 UTC m=+225.061176216 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.108641 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.109312 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-5xv2z" Feb 27 16:26:52 crc kubenswrapper[4814]: E0227 16:26:52.109488 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:52.609402935 +0000 UTC m=+225.062027765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.116694 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-2n4tf" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.122471 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx4gm\" (UniqueName: \"kubernetes.io/projected/3b269e43-ebd6-4f71-8963-669c60763efa-kube-api-access-zx4gm\") pod \"machine-config-controller-84d6567774-g4vb2\" (UID: \"3b269e43-ebd6-4f71-8963-669c60763efa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.133547 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs"] Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.138431 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggb4k\" (UniqueName: \"kubernetes.io/projected/d87f2169-8a70-42de-aaf5-7728ff95fa50-kube-api-access-ggb4k\") pod \"collect-profiles-29536815-pqwsb\" (UID: \"d87f2169-8a70-42de-aaf5-7728ff95fa50\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.144963 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.159902 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nstsm\" (UniqueName: \"kubernetes.io/projected/634e6fbd-87f9-4205-b01d-90be6eb4a015-kube-api-access-nstsm\") pod \"migrator-59844c95c7-twcnw\" (UID: \"634e6fbd-87f9-4205-b01d-90be6eb4a015\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twcnw" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.169083 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.172841 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6lp4\" (UniqueName: \"kubernetes.io/projected/3e0b793d-24a2-4c73-9357-fe193c32f71b-kube-api-access-q6lp4\") pod \"machine-config-operator-74547568cd-w4f4k\" (UID: \"3e0b793d-24a2-4c73-9357-fe193c32f71b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.176588 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twcnw" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.192452 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.194783 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d32d7bef-558f-4dfd-9260-408cb47fc461-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rs9mx\" (UID: \"d32d7bef-558f-4dfd-9260-408cb47fc461\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.200903 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.210372 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:52 crc kubenswrapper[4814]: E0227 16:26:52.210768 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:52.710709643 +0000 UTC m=+225.163334473 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.213280 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m272t\" (UniqueName: \"kubernetes.io/projected/3dafc1d6-a249-4d30-ad93-f065110076c8-kube-api-access-m272t\") pod \"control-plane-machine-set-operator-78cbb6b69f-ws5nw\" (UID: \"3dafc1d6-a249-4d30-ad93-f065110076c8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ws5nw" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.221415 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.228719 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj" Feb 27 16:26:52 crc kubenswrapper[4814]: W0227 16:26:52.242739 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3e3ca4a_f164_4d51_b4a3_6a5f6c5bf09c.slice/crio-76997ccaba6eb1d23570733bfd6f91e77aabc65bde044c8c581b48052bcda5d5 WatchSource:0}: Error finding container 76997ccaba6eb1d23570733bfd6f91e77aabc65bde044c8c581b48052bcda5d5: Status 404 returned error can't find the container with id 76997ccaba6eb1d23570733bfd6f91e77aabc65bde044c8c581b48052bcda5d5 Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.243793 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.244222 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6"] Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.249862 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgcgk\" (UniqueName: \"kubernetes.io/projected/ba54dbcb-efbd-467b-aa4e-994fde5e1c14-kube-api-access-wgcgk\") pod \"multus-admission-controller-857f4d67dd-ncx8v\" (UID: \"ba54dbcb-efbd-467b-aa4e-994fde5e1c14\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ncx8v" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.256191 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld5b5\" (UniqueName: \"kubernetes.io/projected/55b73477-5f70-4dcc-a9aa-3e22ac4efb54-kube-api-access-ld5b5\") pod \"ingress-canary-6hpl2\" (UID: \"55b73477-5f70-4dcc-a9aa-3e22ac4efb54\") " pod="openshift-ingress-canary/ingress-canary-6hpl2" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.265500 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ncx8v" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.280965 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.285037 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7kwh\" (UniqueName: \"kubernetes.io/projected/1a357d8b-0e0e-4ae0-9365-4c4d96316ae2-kube-api-access-p7kwh\") pod \"service-ca-operator-777779d784-4xl77\" (UID: \"1a357d8b-0e0e-4ae0-9365-4c4d96316ae2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xl77" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.289635 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xl77" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.301388 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsmlh\" (UniqueName: \"kubernetes.io/projected/057f89ed-0870-4017-8a0c-a4808d3fb86b-kube-api-access-qsmlh\") pod \"packageserver-d55dfcdfc-8tt9s\" (UID: \"057f89ed-0870-4017-8a0c-a4808d3fb86b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.313863 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jds5c"] Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.314726 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:52 crc kubenswrapper[4814]: E0227 16:26:52.315191 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:52.815174101 +0000 UTC m=+225.267798931 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.340523 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-6hpl2" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.344066 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97d5e809-690b-44bc-9df0-1af46832e149-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lmzk2\" (UID: \"97d5e809-690b-44bc-9df0-1af46832e149\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.346783 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-m9qmb" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.355431 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt2g7\" (UniqueName: \"kubernetes.io/projected/8bc5b690-2e68-4d62-a0cf-349326b8e5e4-kube-api-access-qt2g7\") pod \"olm-operator-6b444d44fb-hl9wj\" (UID: \"8bc5b690-2e68-4d62-a0cf-349326b8e5e4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.360085 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmsmw\" (UniqueName: \"kubernetes.io/projected/dfa49cf2-dc52-4c27-8486-ec1efec61744-kube-api-access-qmsmw\") pod \"marketplace-operator-79b997595-mjlld\" (UID: \"dfa49cf2-dc52-4c27-8486-ec1efec61744\") " pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.378236 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scc9t\" (UniqueName: \"kubernetes.io/projected/0f52f92d-a4bd-41b3-88dd-579a355f6e30-kube-api-access-scc9t\") pod \"csi-hostpathplugin-h8mkp\" (UID: \"0f52f92d-a4bd-41b3-88dd-579a355f6e30\") " pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.408897 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd"] Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.418111 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:52 crc kubenswrapper[4814]: E0227 16:26:52.418992 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:52.918973965 +0000 UTC m=+225.371598805 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.419363 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:52 crc kubenswrapper[4814]: E0227 16:26:52.419661 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:52.919653939 +0000 UTC m=+225.372278779 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.425493 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpbj9\" (UniqueName: \"kubernetes.io/projected/9541f10f-b16a-4dbe-8de2-fb4f381fdb94-kube-api-access-gpbj9\") pod \"kube-storage-version-migrator-operator-b67b599dd-gnk2x\" (UID: \"9541f10f-b16a-4dbe-8de2-fb4f381fdb94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.450179 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ndc8k"] Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.453004 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr6lk\" (UniqueName: \"kubernetes.io/projected/13240f27-5cc2-4e90-9992-24f6beca4212-kube-api-access-lr6lk\") pod \"router-default-5444994796-wv7m2\" (UID: \"13240f27-5cc2-4e90-9992-24f6beca4212\") " pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.457661 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.464069 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.483714 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.510550 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-sbf5t"] Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.510993 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ws5nw" Feb 27 16:26:52 crc kubenswrapper[4814]: W0227 16:26:52.518326 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04649429_b45f_41ff_8777_4e0b321664b9.slice/crio-6dc2778e9848e90ef8d1516da9bd4f5fc2046af783e47d69b1a7de6c764e805c WatchSource:0}: Error finding container 6dc2778e9848e90ef8d1516da9bd4f5fc2046af783e47d69b1a7de6c764e805c: Status 404 returned error can't find the container with id 6dc2778e9848e90ef8d1516da9bd4f5fc2046af783e47d69b1a7de6c764e805c Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.524435 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:52 crc kubenswrapper[4814]: E0227 16:26:52.525159 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:53.025142791 +0000 UTC m=+225.477767621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.537658 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pbsmc"] Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.537884 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.557473 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.573182 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.606543 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54"] Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.620894 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" event={"ID":"2b8fa45e-2136-44a1-9a50-f936b5c22417","Type":"ContainerStarted","Data":"d16902234ca75a2d7614f859b7ae3631debf752938128ebe79c1da001ff140ff"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.620939 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" event={"ID":"2b8fa45e-2136-44a1-9a50-f936b5c22417","Type":"ContainerStarted","Data":"33622f4e01cdbc4268cb0150db92d2acd38068a099177b9f2ff9cc2d5552393d"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.621058 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.628027 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:52 crc kubenswrapper[4814]: E0227 16:26:52.628473 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:53.128456419 +0000 UTC m=+225.581081249 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.629775 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4rvqh" event={"ID":"6bc61e1c-d238-49ae-b344-429bf50959e5","Type":"ContainerStarted","Data":"b9c60146dcfae73963c1b4745c7ada0c5ebcf18f5b07440db875fc74d587cf0a"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.629812 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4rvqh" event={"ID":"6bc61e1c-d238-49ae-b344-429bf50959e5","Type":"ContainerStarted","Data":"d14808c38c30dc072b9c39c62001474ec35f2618e625d593ce98fb2d0f8bc30a"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.629827 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4rvqh" event={"ID":"6bc61e1c-d238-49ae-b344-429bf50959e5","Type":"ContainerStarted","Data":"8341e14fb1fb3d2da44e65f4bdccda13c6819eae9dd427f3733d5b1ba39dbabf"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.633955 4814 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-gmfbn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.634029 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" podUID="2b8fa45e-2136-44a1-9a50-f936b5c22417" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.642752 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" event={"ID":"e245069f-c15d-4ad4-9709-6bec4a52c8c8","Type":"ContainerStarted","Data":"9b7d09b468408fa0f9ff497c12e9a9f37c540443c87af8082ec7d0a8efc78828"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.646467 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" event={"ID":"29babfee-4ec4-4828-90c6-caf8fb0647b2","Type":"ContainerStarted","Data":"42444ba34a81c8422492e20f74d32da4e017ea620658d63ced0e5357bb2b66da"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.646492 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" event={"ID":"29babfee-4ec4-4828-90c6-caf8fb0647b2","Type":"ContainerStarted","Data":"b9e500879d9a16a34a4f3ace10ff7d4572ba0732553f5cc9f0828da0215475fd"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.646502 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" event={"ID":"29babfee-4ec4-4828-90c6-caf8fb0647b2","Type":"ContainerStarted","Data":"cf00bfbcfdff6c29605896888c870bd2c4c376a1036dc8186cd1c3d9b23f1aa0"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.650099 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" event={"ID":"1000b998-ffa1-40c5-8563-0aeb70d172a9","Type":"ContainerStarted","Data":"46c85e5d63ee25a1bc79a3025637a1a712251b11aea6dbf1b31bd48ca586c989"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.650121 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" event={"ID":"1000b998-ffa1-40c5-8563-0aeb70d172a9","Type":"ContainerStarted","Data":"1deaf117186b7d1efa4ddcd141c77174ac58f9bca95b24a5d9fad1354e272549"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.650413 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.657758 4814 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-4m8vp container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.657811 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" podUID="1000b998-ffa1-40c5-8563-0aeb70d172a9" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.679114 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" event={"ID":"46c9f2b1-15f1-418b-a086-5ea9972185a0","Type":"ContainerStarted","Data":"8f5d20967bc67a22490a65cf2be6f421c9a411798a496ecd96a4f36728d11d67"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.679185 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" event={"ID":"46c9f2b1-15f1-418b-a086-5ea9972185a0","Type":"ContainerStarted","Data":"c31c6b8b0a1b4a7b9cc15bb7eed86c317746d48533b2da372a2c4c5d0e79939d"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.684944 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" event={"ID":"ec601e1d-d940-4bbb-9cce-83be390511ef","Type":"ContainerStarted","Data":"040a0f87d4dd9b296f321af2cb4a9f8f1f9afd14b01d467090cbdde0409be8d8"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.684994 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" event={"ID":"ec601e1d-d940-4bbb-9cce-83be390511ef","Type":"ContainerStarted","Data":"1f75321f992945c15bb021b573b0ff5038341da7671b291376c15cd482a18aa6"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.685300 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.691915 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" event={"ID":"bc099c2d-2785-4509-bd8b-29469341f56d","Type":"ContainerStarted","Data":"5106a7a11b69d63039ca8d767b7d92bad7bdbd7acc56932f8206e8a9eb375c3d"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.701408 4814 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-k4p69 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.19:6443/healthz\": dial tcp 10.217.0.19:6443: connect: connection refused" start-of-body= Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.701474 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" podUID="ec601e1d-d940-4bbb-9cce-83be390511ef" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.19:6443/healthz\": dial tcp 10.217.0.19:6443: connect: connection refused" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.703026 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" event={"ID":"2dc05d09-1fba-4a2e-93fe-3db86e8ff333","Type":"ContainerStarted","Data":"b158570134b3672adcff998106567b6daf700d2e95df9b194ac0fcdd7c4d2bba"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.703953 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" event={"ID":"04649429-b45f-41ff-8777-4e0b321664b9","Type":"ContainerStarted","Data":"6dc2778e9848e90ef8d1516da9bd4f5fc2046af783e47d69b1a7de6c764e805c"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.713726 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" event={"ID":"8b44d028-0edc-4477-9f01-236c6d59d9e8","Type":"ContainerStarted","Data":"8b8fdcb0889649fd6e910fa2014324655a98bce5d70dd7db32fcd9f5dcd74dee"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.717484 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7x2qp" event={"ID":"f487b7f4-c24b-431e-8868-a0285c8ed7bc","Type":"ContainerStarted","Data":"6c107d6ef30a6b7c63776b99a6e0b9b6c74689983e810164c431b4553ef4f089"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.717528 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7x2qp" event={"ID":"f487b7f4-c24b-431e-8868-a0285c8ed7bc","Type":"ContainerStarted","Data":"e461de5943467cdbf6f7248a16c5dfd1c11ff1400fc8140953caf4be6f92c322"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.721415 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs" event={"ID":"d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c","Type":"ContainerStarted","Data":"76997ccaba6eb1d23570733bfd6f91e77aabc65bde044c8c581b48052bcda5d5"} Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.729123 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:52 crc kubenswrapper[4814]: E0227 16:26:52.729247 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:53.229228929 +0000 UTC m=+225.681853759 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.729594 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:52 crc kubenswrapper[4814]: E0227 16:26:52.732830 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:53.232820763 +0000 UTC m=+225.685445593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.755154 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.796441 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-qzwns"] Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.798852 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536826-dgwn4"] Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.831843 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:52 crc kubenswrapper[4814]: E0227 16:26:52.832492 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:53.332462323 +0000 UTC m=+225.785087153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.833313 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:52 crc kubenswrapper[4814]: E0227 16:26:52.836573 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:53.336565325 +0000 UTC m=+225.789190155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.925732 4814 ???:1] "http: TLS handshake error from 192.168.126.11:56226: no serving certificate available for the kubelet" Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.935082 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:52 crc kubenswrapper[4814]: E0227 16:26:52.940305 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:53.440237285 +0000 UTC m=+225.892862125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.940545 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:52 crc kubenswrapper[4814]: E0227 16:26:52.941387 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:53.441368064 +0000 UTC m=+225.893992894 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:52 crc kubenswrapper[4814]: I0227 16:26:52.950610 4814 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.018443 4814 ???:1] "http: TLS handshake error from 192.168.126.11:56232: no serving certificate available for the kubelet" Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.042428 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:53 crc kubenswrapper[4814]: E0227 16:26:53.042788 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:53.542765415 +0000 UTC m=+225.995390245 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.119122 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" podStartSLOduration=160.119092041 podStartE2EDuration="2m40.119092041s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:53.114871405 +0000 UTC m=+225.567496235" watchObservedRunningTime="2026-02-27 16:26:53.119092041 +0000 UTC m=+225.571716871" Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.122396 4814 ???:1] "http: TLS handshake error from 192.168.126.11:56248: no serving certificate available for the kubelet" Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.144663 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:53 crc kubenswrapper[4814]: E0227 16:26:53.145152 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:53.6451363 +0000 UTC m=+226.097761130 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.221793 4814 ???:1] "http: TLS handshake error from 192.168.126.11:56262: no serving certificate available for the kubelet" Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.245887 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:53 crc kubenswrapper[4814]: E0227 16:26:53.246212 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:53.74619912 +0000 UTC m=+226.198823950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.328538 4814 ???:1] "http: TLS handshake error from 192.168.126.11:56276: no serving certificate available for the kubelet" Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.350936 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:53 crc kubenswrapper[4814]: E0227 16:26:53.351346 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:53.851328551 +0000 UTC m=+226.303953381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.428277 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5xv2z"] Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.429182 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.435306 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-2n4tf"] Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.449854 4814 ???:1] "http: TLS handshake error from 192.168.126.11:56290: no serving certificate available for the kubelet" Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.452932 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:53 crc kubenswrapper[4814]: E0227 16:26:53.453449 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:53.953426206 +0000 UTC m=+226.406051036 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.554925 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.556624 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4rvqh" podStartSLOduration=160.556591158 podStartE2EDuration="2m40.556591158s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:53.547011398 +0000 UTC m=+225.999636228" watchObservedRunningTime="2026-02-27 16:26:53.556591158 +0000 UTC m=+226.009215998" Feb 27 16:26:53 crc kubenswrapper[4814]: E0227 16:26:53.557593 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:54.057575532 +0000 UTC m=+226.510200362 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.622643 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rzv4n" podStartSLOduration=160.622624679 podStartE2EDuration="2m40.622624679s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:53.621307043 +0000 UTC m=+226.073931873" watchObservedRunningTime="2026-02-27 16:26:53.622624679 +0000 UTC m=+226.075249509" Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.660460 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:53 crc kubenswrapper[4814]: E0227 16:26:53.661215 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:54.161199501 +0000 UTC m=+226.613824321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.683210 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-9vmj5" podStartSLOduration=160.68319214 podStartE2EDuration="2m40.68319214s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:53.6817379 +0000 UTC m=+226.134362790" watchObservedRunningTime="2026-02-27 16:26:53.68319214 +0000 UTC m=+226.135816970" Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.753839 4814 ???:1] "http: TLS handshake error from 192.168.126.11:56302: no serving certificate available for the kubelet" Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.753874 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs" event={"ID":"d3e3ca4a-f164-4d51-b4a3-6a5f6c5bf09c","Type":"ContainerStarted","Data":"c471bacd01cdfa9359ba6b5c52a297bcb24e64d622256d7df0c9d08ee7f08c43"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.755750 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" event={"ID":"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b","Type":"ContainerStarted","Data":"bd582824aa4142901ca4a48fa2cf2c0a60e6e289ba6528aa13807ec956c24298"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.756655 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-m9qmb" event={"ID":"d518d443-b142-4235-8720-31d8f421c794","Type":"ContainerStarted","Data":"247efb3e66ce871b02b663b06d93863c35bfb8fa67adbc6d69ec3d0253d19179"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.756680 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-m9qmb" event={"ID":"d518d443-b142-4235-8720-31d8f421c794","Type":"ContainerStarted","Data":"c221e2764b89ac96caa7641a4d2d62d4260582f768ac0bb694db2989d7078a4d"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.764270 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:53 crc kubenswrapper[4814]: E0227 16:26:53.764645 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:54.264634292 +0000 UTC m=+226.717259122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.767304 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2n4tf" event={"ID":"98112972-5121-4d93-8490-5156a1a2f3ca","Type":"ContainerStarted","Data":"69b05ad8ec3c42f187b3e8da54000377f14fc59f82f06a2592e96fb3ee26d0a2"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.779064 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-qzwns" event={"ID":"8676c139-d437-4f7a-9234-de7c8a1b8dd1","Type":"ContainerStarted","Data":"2a8ec489aa3b44fafae5f266c4d51e3950f0034e6d767e86b07a44125e90d8f8"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.779113 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-qzwns" event={"ID":"8676c139-d437-4f7a-9234-de7c8a1b8dd1","Type":"ContainerStarted","Data":"14f330c144183112d6f1466f9d68737d24019a390aa3408ceea32ab1264cdc0b"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.780329 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-sbf5t" event={"ID":"9b78e04c-f222-4aae-be52-85abd4ecf391","Type":"ContainerStarted","Data":"8cb92a6c66e045a3d8132e7cfa0f449ea0561059ea0428b80cddcd76329af1ff"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.781920 4814 generic.go:334] "Generic (PLEG): container finished" podID="2dc05d09-1fba-4a2e-93fe-3db86e8ff333" containerID="e1e9b3cc2bba84dab5536ae47cf5d0b2ca71df9c60eb5b31d1084713827dc4eb" exitCode=0 Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.781986 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" event={"ID":"2dc05d09-1fba-4a2e-93fe-3db86e8ff333","Type":"ContainerDied","Data":"e1e9b3cc2bba84dab5536ae47cf5d0b2ca71df9c60eb5b31d1084713827dc4eb"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.783368 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" event={"ID":"04649429-b45f-41ff-8777-4e0b321664b9","Type":"ContainerStarted","Data":"91d9e243310c918acb6eefbb8ce93149c7806deb5dba390cef8aecb6ca6ecc47"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.787815 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-wv7m2" event={"ID":"13240f27-5cc2-4e90-9992-24f6beca4212","Type":"ContainerStarted","Data":"43a2d7e39245011ef2fa1e8599cd087318447835e8a16171c41c7872806eda69"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.787870 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-wv7m2" event={"ID":"13240f27-5cc2-4e90-9992-24f6beca4212","Type":"ContainerStarted","Data":"6deb3546546182e139a1a36626c940b03c60d27aafa718624b44d4d13cf48306"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.797099 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536826-dgwn4" event={"ID":"95913cb4-39f1-44c0-ac49-0a2d51047679","Type":"ContainerStarted","Data":"563fc3d178b9caef3700f56e12565007094ababa2341ca39792590befcbab569"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.798421 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" event={"ID":"e245069f-c15d-4ad4-9709-6bec4a52c8c8","Type":"ContainerStarted","Data":"a8c493fe011a1df84f1560b32bc355f51669b948dd4e26c0d60d2f640719be36"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.802230 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-5xv2z" event={"ID":"db53a48c-000d-4e92-929e-6c6bf561119a","Type":"ContainerStarted","Data":"f631547c5eda0a8bd32737f1b0b629b1133f5c1277eda2603d1df5f8bff92b77"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.803144 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-7x2qp" podStartSLOduration=160.803131941 podStartE2EDuration="2m40.803131941s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:53.80249292 +0000 UTC m=+226.255117750" watchObservedRunningTime="2026-02-27 16:26:53.803131941 +0000 UTC m=+226.255756771" Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.807446 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" event={"ID":"bc099c2d-2785-4509-bd8b-29469341f56d","Type":"ContainerStarted","Data":"78785355536f63ca5ba6003f157168617b76d7f1fa8ff04859529850560062c2"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.815879 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54" event={"ID":"008442d1-8ae3-4669-945d-d7f3389855a2","Type":"ContainerStarted","Data":"f8934d831f9463dbdcca4cb70659a2a1468cac65c2233f2409da9ddb83b1ce87"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.815940 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54" event={"ID":"008442d1-8ae3-4669-945d-d7f3389855a2","Type":"ContainerStarted","Data":"f0d28d3fdace722026a42a46cb61658d7352dbeb00feb0efebf6e57ee1fe0053"} Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.833877 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.840356 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" podStartSLOduration=160.840323316 podStartE2EDuration="2m40.840323316s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:53.840244003 +0000 UTC m=+226.292868833" watchObservedRunningTime="2026-02-27 16:26:53.840323316 +0000 UTC m=+226.292948146" Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.868608 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:53 crc kubenswrapper[4814]: E0227 16:26:53.884808 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:54.384784301 +0000 UTC m=+226.837409131 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.974887 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:53 crc kubenswrapper[4814]: I0227 16:26:53.980100 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:53 crc kubenswrapper[4814]: E0227 16:26:53.980491 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:54.480470285 +0000 UTC m=+226.933095115 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.024176 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktzsn" podStartSLOduration=161.023952717 podStartE2EDuration="2m41.023952717s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:54.021714499 +0000 UTC m=+226.474339319" watchObservedRunningTime="2026-02-27 16:26:54.023952717 +0000 UTC m=+226.476577567" Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.076268 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:54 crc kubenswrapper[4814]: E0227 16:26:54.076847 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:54.576826453 +0000 UTC m=+227.029451283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.135715 4814 ???:1] "http: TLS handshake error from 192.168.126.11:56314: no serving certificate available for the kubelet" Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.135795 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pwdrv"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.143477 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.178490 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:54 crc kubenswrapper[4814]: E0227 16:26:54.179149 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:54.679134356 +0000 UTC m=+227.131759186 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.184341 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.195508 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.195613 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" podStartSLOduration=160.195603944 podStartE2EDuration="2m40.195603944s" podCreationTimestamp="2026-02-27 16:24:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:54.149201132 +0000 UTC m=+226.601825962" watchObservedRunningTime="2026-02-27 16:26:54.195603944 +0000 UTC m=+226.648228774" Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.250394 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4xl77"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.271150 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.281000 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:54 crc kubenswrapper[4814]: E0227 16:26:54.281472 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:54.781447029 +0000 UTC m=+227.234071859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.286928 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ncx8v"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.311183 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.383352 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:54 crc kubenswrapper[4814]: E0227 16:26:54.383855 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:54.883834634 +0000 UTC m=+227.336459464 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.393486 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-twcnw"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.438114 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xx5hs" podStartSLOduration=161.438090478 podStartE2EDuration="2m41.438090478s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:54.407777181 +0000 UTC m=+226.860402011" watchObservedRunningTime="2026-02-27 16:26:54.438090478 +0000 UTC m=+226.890715308" Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.484567 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:54 crc kubenswrapper[4814]: E0227 16:26:54.484938 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:54.984921095 +0000 UTC m=+227.437545925 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.485923 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.523093 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-jds5c" podStartSLOduration=161.523062582 podStartE2EDuration="2m41.523062582s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:54.514592959 +0000 UTC m=+226.967217799" watchObservedRunningTime="2026-02-27 16:26:54.523062582 +0000 UTC m=+226.975687412" Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.552029 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.556849 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.556909 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-6hpl2"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.587912 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:54 crc kubenswrapper[4814]: E0227 16:26:54.592396 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:55.092376345 +0000 UTC m=+227.545001175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.595718 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ws5nw"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.606825 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.635873 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6c54" podStartSLOduration=161.635837436 podStartE2EDuration="2m41.635837436s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:54.612426547 +0000 UTC m=+227.065051377" watchObservedRunningTime="2026-02-27 16:26:54.635837436 +0000 UTC m=+227.088462256" Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.638906 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.670561 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-qzwns" podStartSLOduration=160.670539815 podStartE2EDuration="2m40.670539815s" podCreationTimestamp="2026-02-27 16:24:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:54.667651375 +0000 UTC m=+227.120276205" watchObservedRunningTime="2026-02-27 16:26:54.670539815 +0000 UTC m=+227.123164645" Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.689124 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:54 crc kubenswrapper[4814]: E0227 16:26:54.689592 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:55.189568172 +0000 UTC m=+227.642193002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:54 crc kubenswrapper[4814]: W0227 16:26:54.698666 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97d5e809_690b_44bc_9df0_1af46832e149.slice/crio-46381c1e1051dd76a2447bd0d51f0868f3a3e4d42e6cdda7627fba1e93efa49b WatchSource:0}: Error finding container 46381c1e1051dd76a2447bd0d51f0868f3a3e4d42e6cdda7627fba1e93efa49b: Status 404 returned error can't find the container with id 46381c1e1051dd76a2447bd0d51f0868f3a3e4d42e6cdda7627fba1e93efa49b Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.739822 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-wv7m2" podStartSLOduration=161.739793956 podStartE2EDuration="2m41.739793956s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:54.724712215 +0000 UTC m=+227.177337045" watchObservedRunningTime="2026-02-27 16:26:54.739793956 +0000 UTC m=+227.192418786" Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.785061 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.791829 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:54 crc kubenswrapper[4814]: E0227 16:26:54.792407 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:55.292391962 +0000 UTC m=+227.745016802 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.798960 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mjlld"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.800536 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:26:54 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:26:54 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:26:54 crc kubenswrapper[4814]: healthz check failed Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.800588 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.800604 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-m9qmb" podStartSLOduration=5.8005807350000005 podStartE2EDuration="5.800580735s" podCreationTimestamp="2026-02-27 16:26:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:54.779827159 +0000 UTC m=+227.232451989" watchObservedRunningTime="2026-02-27 16:26:54.800580735 +0000 UTC m=+227.253205585" Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.876395 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-h8mkp"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.880519 4814 ???:1] "http: TLS handshake error from 192.168.126.11:56318: no serving certificate available for the kubelet" Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.913472 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:54 crc kubenswrapper[4814]: E0227 16:26:54.913946 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:55.413917059 +0000 UTC m=+227.866541879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.916642 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:54 crc kubenswrapper[4814]: E0227 16:26:54.917010 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:55.416998935 +0000 UTC m=+227.869623765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.915311 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" event={"ID":"2dc05d09-1fba-4a2e-93fe-3db86e8ff333","Type":"ContainerStarted","Data":"20982f4a4fec38ae8b6182c9ab224c0797823173fa6605a58eee410e2ecfaa47"} Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.925440 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.945497 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.947384 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj"] Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.949797 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2" event={"ID":"97d5e809-690b-44bc-9df0-1af46832e149","Type":"ContainerStarted","Data":"46381c1e1051dd76a2447bd0d51f0868f3a3e4d42e6cdda7627fba1e93efa49b"} Feb 27 16:26:54 crc kubenswrapper[4814]: I0227 16:26:54.966148 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-sbf5t" event={"ID":"9b78e04c-f222-4aae-be52-85abd4ecf391","Type":"ContainerStarted","Data":"accc6b8e6add45ff792ff0455ddfd7f8759f39cc958a1e1f82ca9b75e6769727"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.010863 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj" event={"ID":"5768bed9-9187-4bef-ae5e-bc003793d741","Type":"ContainerStarted","Data":"87068e69092ef3366275647c3b3de270d02e77f8a15054222a29e031bae8565e"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.010908 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj" event={"ID":"5768bed9-9187-4bef-ae5e-bc003793d741","Type":"ContainerStarted","Data":"4d72edb02178de701bb83b2c0037c236bb04060cf38a290724d3a93dc76a0b27"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.018418 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:55 crc kubenswrapper[4814]: E0227 16:26:55.019197 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:55.519162173 +0000 UTC m=+227.971787003 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.025463 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2n4tf" event={"ID":"98112972-5121-4d93-8490-5156a1a2f3ca","Type":"ContainerStarted","Data":"a5601efde7b5290568bdb95cb45e2f1887870286a9012d3680584eb87088aa10"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.026002 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-2n4tf" Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.028717 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-6hpl2" event={"ID":"55b73477-5f70-4dcc-a9aa-3e22ac4efb54","Type":"ContainerStarted","Data":"dc43897a325cfad30f75f2013eeec2046d220f3323f49e46de1c8b9ffdf9146c"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.033116 4814 patch_prober.go:28] interesting pod/downloads-7954f5f757-2n4tf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.033197 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2n4tf" podUID="98112972-5121-4d93-8490-5156a1a2f3ca" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.035523 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pwdrv" event={"ID":"fc590a1b-ba51-421c-a089-fb0def054fb2","Type":"ContainerStarted","Data":"b0edcf6be319a716d68fb36cb5e78c271ccbfdeb46dfdab148e57c687d14e960"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.038741 4814 generic.go:334] "Generic (PLEG): container finished" podID="04649429-b45f-41ff-8777-4e0b321664b9" containerID="91d9e243310c918acb6eefbb8ce93149c7806deb5dba390cef8aecb6ca6ecc47" exitCode=0 Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.038855 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" event={"ID":"04649429-b45f-41ff-8777-4e0b321664b9","Type":"ContainerDied","Data":"91d9e243310c918acb6eefbb8ce93149c7806deb5dba390cef8aecb6ca6ecc47"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.038933 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.038946 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" event={"ID":"04649429-b45f-41ff-8777-4e0b321664b9","Type":"ContainerStarted","Data":"beca7fa740c6ad919d99b437edafe08031b86bbeb56b3897714efe196b562d40"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.044744 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x" event={"ID":"9541f10f-b16a-4dbe-8de2-fb4f381fdb94","Type":"ContainerStarted","Data":"3a929bc4c58ec7a90524f4759265cd96ecf1799d87a63259c3e1882591b814e4"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.047543 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ws5nw" event={"ID":"3dafc1d6-a249-4d30-ad93-f065110076c8","Type":"ContainerStarted","Data":"ea7f464b9153e398cee0fe628785f64f7d81ec5faa5f92585eb8d839bd5c8d6f"} Feb 27 16:26:55 crc kubenswrapper[4814]: W0227 16:26:55.061795 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd32d7bef_558f_4dfd_9260_408cb47fc461.slice/crio-a986f47015585145d422feaf718673153d2c306d8f22ff311830b626cb4546ec WatchSource:0}: Error finding container a986f47015585145d422feaf718673153d2c306d8f22ff311830b626cb4546ec: Status 404 returned error can't find the container with id a986f47015585145d422feaf718673153d2c306d8f22ff311830b626cb4546ec Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.085001 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-5xv2z" event={"ID":"db53a48c-000d-4e92-929e-6c6bf561119a","Type":"ContainerStarted","Data":"938bc841c6f286d3243992ce263260f489f95da119c5fe5cfcd559d2b6aca219"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.085298 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-5xv2z" Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.092084 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" podStartSLOduration=161.09206553 podStartE2EDuration="2m41.09206553s" podCreationTimestamp="2026-02-27 16:24:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:55.090513147 +0000 UTC m=+227.543137977" watchObservedRunningTime="2026-02-27 16:26:55.09206553 +0000 UTC m=+227.544690360" Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.096554 4814 patch_prober.go:28] interesting pod/console-operator-58897d9998-5xv2z container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.096647 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-5xv2z" podUID="db53a48c-000d-4e92-929e-6c6bf561119a" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.119712 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.121130 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-5xv2z" podStartSLOduration=162.121113184 podStartE2EDuration="2m42.121113184s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:55.114879268 +0000 UTC m=+227.567504098" watchObservedRunningTime="2026-02-27 16:26:55.121113184 +0000 UTC m=+227.573738014" Feb 27 16:26:55 crc kubenswrapper[4814]: E0227 16:26:55.121976 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:55.621964002 +0000 UTC m=+228.074588832 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.139835 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" event={"ID":"bc099c2d-2785-4509-bd8b-29469341f56d","Type":"ContainerStarted","Data":"f93472519c04c8a8606d7391534eb7ea0d90767fcf9b62b103968e28aabe1bb7"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.156166 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-2n4tf" podStartSLOduration=162.156148933 podStartE2EDuration="2m42.156148933s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:55.155872433 +0000 UTC m=+227.608497263" watchObservedRunningTime="2026-02-27 16:26:55.156148933 +0000 UTC m=+227.608773763" Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.192984 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" podStartSLOduration=162.192963224 podStartE2EDuration="2m42.192963224s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:55.192221358 +0000 UTC m=+227.644846188" watchObservedRunningTime="2026-02-27 16:26:55.192963224 +0000 UTC m=+227.645588054" Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.202915 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ncx8v" event={"ID":"ba54dbcb-efbd-467b-aa4e-994fde5e1c14","Type":"ContainerStarted","Data":"2fd2360f74acd7d26e3d4480c823762f1452b1f3e6f2f65a22c10c7e98632645"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.215775 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" event={"ID":"6401ae31-5292-4d4f-8d39-a5682c0d38ee","Type":"ContainerStarted","Data":"050c1168839c0864011bd5da0b63d092197b678ba307344b9a6625e41ba2f20f"} Feb 27 16:26:55 crc kubenswrapper[4814]: E0227 16:26:55.221109 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:55.721035914 +0000 UTC m=+228.173660744 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.221927 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.222322 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:55 crc kubenswrapper[4814]: E0227 16:26:55.223924 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:55.723856541 +0000 UTC m=+228.176481371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.242327 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" event={"ID":"3e0b793d-24a2-4c73-9357-fe193c32f71b","Type":"ContainerStarted","Data":"067ae27eabcc4c803d264b3afb7555de78b8c33b6c6c896e0f46442dda30004c"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.257782 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-ndc8k" podStartSLOduration=162.257762872 podStartE2EDuration="2m42.257762872s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:55.255332298 +0000 UTC m=+227.707957138" watchObservedRunningTime="2026-02-27 16:26:55.257762872 +0000 UTC m=+227.710387702" Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.270522 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twcnw" event={"ID":"634e6fbd-87f9-4205-b01d-90be6eb4a015","Type":"ContainerStarted","Data":"e822cae7fec29065c6e5920e1f0c0aeb842bdc38c9e084d0b018a5414928e4ba"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.274271 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n" event={"ID":"d367fc6a-d4c4-4e63-81da-c547e07eea8d","Type":"ContainerStarted","Data":"1ce49bca165b018456ee3d11567134a932af5f495a9e990b463140758f4ce403"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.294342 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2" event={"ID":"3b269e43-ebd6-4f71-8963-669c60763efa","Type":"ContainerStarted","Data":"2a11c11624b7b92a7e8cfc5ce23af6315a51aa83de8b48446cf2d00c21bbb012"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.294427 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2" event={"ID":"3b269e43-ebd6-4f71-8963-669c60763efa","Type":"ContainerStarted","Data":"43dedacf6391c0e83ba6139c5d2a0dec709c91c4b6220a0be34c9d940ee9b4d8"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.317901 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xl77" event={"ID":"1a357d8b-0e0e-4ae0-9365-4c4d96316ae2","Type":"ContainerStarted","Data":"dfece27d5164bbb6c0ceef0e919e12197f28ec5a42a87e3d1cda2223bba820ef"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.321834 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" event={"ID":"d87f2169-8a70-42de-aaf5-7728ff95fa50","Type":"ContainerStarted","Data":"ae16799b7d5aaebd0d3a0366d13149ce400a26bbfd718fc2dd9891cab15194dd"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.323668 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:55 crc kubenswrapper[4814]: E0227 16:26:55.324068 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:55.824048301 +0000 UTC m=+228.276673131 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.325647 4814 generic.go:334] "Generic (PLEG): container finished" podID="e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b" containerID="06475eb07864c49100912b86a7e77aa5d6669588e7a6403cd9d4e57dffc7ab97" exitCode=0 Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.325706 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" event={"ID":"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b","Type":"ContainerDied","Data":"06475eb07864c49100912b86a7e77aa5d6669588e7a6403cd9d4e57dffc7ab97"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.349367 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xl77" podStartSLOduration=161.349348934 podStartE2EDuration="2m41.349348934s" podCreationTimestamp="2026-02-27 16:24:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:55.348436393 +0000 UTC m=+227.801061223" watchObservedRunningTime="2026-02-27 16:26:55.349348934 +0000 UTC m=+227.801973764" Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.379637 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" event={"ID":"cdc3c78e-177f-4fcd-ab8d-c333497848aa","Type":"ContainerStarted","Data":"55015c66ba3bf9b19dabcae29fb43bbfcc3e7e5cd1c6a5bbe477d262bb5ff678"} Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.431730 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:55 crc kubenswrapper[4814]: E0227 16:26:55.432157 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:55.932132194 +0000 UTC m=+228.384757024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.479381 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" podStartSLOduration=162.479341924 podStartE2EDuration="2m42.479341924s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:55.464509881 +0000 UTC m=+227.917134711" watchObservedRunningTime="2026-02-27 16:26:55.479341924 +0000 UTC m=+227.931966754" Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.532603 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:55 crc kubenswrapper[4814]: E0227 16:26:55.532722 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:56.032705746 +0000 UTC m=+228.485330576 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.532994 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:55 crc kubenswrapper[4814]: E0227 16:26:55.536093 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:56.036082602 +0000 UTC m=+228.488707432 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.634010 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:55 crc kubenswrapper[4814]: E0227 16:26:55.634332 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:56.134310144 +0000 UTC m=+228.586934974 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.634502 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:55 crc kubenswrapper[4814]: E0227 16:26:55.634804 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:56.134796311 +0000 UTC m=+228.587421131 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.755089 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:55 crc kubenswrapper[4814]: E0227 16:26:55.755944 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:56.255923474 +0000 UTC m=+228.708548294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.802520 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:26:55 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:26:55 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:26:55 crc kubenswrapper[4814]: healthz check failed Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.802588 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.857624 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:55 crc kubenswrapper[4814]: E0227 16:26:55.858213 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:56.358200455 +0000 UTC m=+228.810825285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.961007 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:55 crc kubenswrapper[4814]: E0227 16:26:55.961279 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:56.461237174 +0000 UTC m=+228.913861994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:55 crc kubenswrapper[4814]: I0227 16:26:55.962222 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:55 crc kubenswrapper[4814]: E0227 16:26:55.962547 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:56.462535849 +0000 UTC m=+228.915160679 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.064559 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:56 crc kubenswrapper[4814]: E0227 16:26:56.064909 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:56.564872603 +0000 UTC m=+229.017497583 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.065204 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:56 crc kubenswrapper[4814]: E0227 16:26:56.066086 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:56.566068674 +0000 UTC m=+229.018693504 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.166792 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:56 crc kubenswrapper[4814]: E0227 16:26:56.168359 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:56.668340826 +0000 UTC m=+229.120965656 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.217750 4814 ???:1] "http: TLS handshake error from 192.168.126.11:56326: no serving certificate available for the kubelet" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.268999 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:56 crc kubenswrapper[4814]: E0227 16:26:56.269348 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:56.769335072 +0000 UTC m=+229.221959902 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.270416 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4m8vp"] Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.320950 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn"] Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.370077 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:56 crc kubenswrapper[4814]: E0227 16:26:56.370660 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:56.870640331 +0000 UTC m=+229.323265171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.414454 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" event={"ID":"3e0b793d-24a2-4c73-9357-fe193c32f71b","Type":"ContainerStarted","Data":"09c10b0ffb6e604688d2fc8f8376d603db4f07b07198043195c4c84695dad6fa"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.414500 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" event={"ID":"3e0b793d-24a2-4c73-9357-fe193c32f71b","Type":"ContainerStarted","Data":"db9add9cbc55f3327dbef353a0292c20fe3ed72a94e77c27965068b3c82a63a0"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.418804 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" event={"ID":"d87f2169-8a70-42de-aaf5-7728ff95fa50","Type":"ContainerStarted","Data":"b86ccc9ec3b77950f4a33f3517ef3218c0cf88bf4a5d8200fb3cda18db954532"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.437677 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" event={"ID":"8bc5b690-2e68-4d62-a0cf-349326b8e5e4","Type":"ContainerStarted","Data":"904bbfb25385c97e79248f5f4fe65845ed98e4177da390da7bb4b71e410f0403"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.437738 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" event={"ID":"8bc5b690-2e68-4d62-a0cf-349326b8e5e4","Type":"ContainerStarted","Data":"fddf85a750d66c5711d4b7d9f044e14b5ff8c0b494c8601db6681f0dcac44cc0"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.438295 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.446066 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pwdrv" event={"ID":"fc590a1b-ba51-421c-a089-fb0def054fb2","Type":"ContainerStarted","Data":"50e41492ef47a87f36a9eae2938c77a85d317c3611b21a826adf968ea821ec56"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.447051 4814 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-hl9wj container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.447093 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" podUID="8bc5b690-2e68-4d62-a0cf-349326b8e5e4" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.459080 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-sbf5t" event={"ID":"9b78e04c-f222-4aae-be52-85abd4ecf391","Type":"ContainerStarted","Data":"bd94704bfe8b6a9c72ee9368498c1d82134e95cc6e91d9da9cfa8c30cf69c01c"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.459300 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-sbf5t" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.471316 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:56 crc kubenswrapper[4814]: E0227 16:26:56.471637 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:56.971624048 +0000 UTC m=+229.424248878 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.498791 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ws5nw" event={"ID":"3dafc1d6-a249-4d30-ad93-f065110076c8","Type":"ContainerStarted","Data":"dd005383809fe066d36e652584ec88d3fd73b484c9067b6804c362731f7cf6e4"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.505767 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x" event={"ID":"9541f10f-b16a-4dbe-8de2-fb4f381fdb94","Type":"ContainerStarted","Data":"7d2df2b276f49b9172a50bfa4acdbc03d54bc6bac8ac370fd00b5ced6a2b8fb6"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.508691 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx" event={"ID":"d32d7bef-558f-4dfd-9260-408cb47fc461","Type":"ContainerStarted","Data":"a986f47015585145d422feaf718673153d2c306d8f22ff311830b626cb4546ec"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.517430 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2" event={"ID":"97d5e809-690b-44bc-9df0-1af46832e149","Type":"ContainerStarted","Data":"b064af6ea956f1fd7af6482fb972fb66388efd7b139ba5f38031908e2dc2d125"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.525271 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" podStartSLOduration=163.52524335 podStartE2EDuration="2m43.52524335s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:56.524659289 +0000 UTC m=+228.977284119" watchObservedRunningTime="2026-02-27 16:26:56.52524335 +0000 UTC m=+228.977868180" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.526666 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w4f4k" podStartSLOduration=163.526658749 podStartE2EDuration="2m43.526658749s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:56.478473545 +0000 UTC m=+228.931098375" watchObservedRunningTime="2026-02-27 16:26:56.526658749 +0000 UTC m=+228.979283579" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.531099 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n" event={"ID":"d367fc6a-d4c4-4e63-81da-c547e07eea8d","Type":"ContainerStarted","Data":"a2a4fd00b335c1202668b242b4222243e027b4b7aaf4131011376c0e8a2e9be4"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.535576 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xl77" event={"ID":"1a357d8b-0e0e-4ae0-9365-4c4d96316ae2","Type":"ContainerStarted","Data":"1fb99ce9a833ee85421b87aae4f32a745b053fe036c9aa416e7d58efab557e40"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.557023 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" event={"ID":"0f52f92d-a4bd-41b3-88dd-579a355f6e30","Type":"ContainerStarted","Data":"12215135f0f06e693020ce937f3190217fb2813a7881553f47cf5be1a8db4299"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.567857 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ncx8v" event={"ID":"ba54dbcb-efbd-467b-aa4e-994fde5e1c14","Type":"ContainerStarted","Data":"345329e4afba305e797ddd257ec040d1cad1280b9935f0ab8c08fb5131a72d79"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.572981 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:56 crc kubenswrapper[4814]: E0227 16:26:56.574561 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:57.074544672 +0000 UTC m=+229.527169502 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.584596 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" event={"ID":"dfa49cf2-dc52-4c27-8486-ec1efec61744","Type":"ContainerStarted","Data":"c368bbd9a1ae5371c62ca4d7ecc6b992fe4809aa77ddb9bd665461038649f222"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.584843 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" event={"ID":"dfa49cf2-dc52-4c27-8486-ec1efec61744","Type":"ContainerStarted","Data":"436c02e76e48674afb5902acd48631763f9a4ab4ccd2d5dc2c880d50a8814641"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.585909 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.594154 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" podStartSLOduration=163.594139228 podStartE2EDuration="2m43.594139228s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:56.575521826 +0000 UTC m=+229.028146656" watchObservedRunningTime="2026-02-27 16:26:56.594139228 +0000 UTC m=+229.046764058" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.600423 4814 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mjlld container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.600486 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" podUID="dfa49cf2-dc52-4c27-8486-ec1efec61744" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.601055 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nh2sn" event={"ID":"cdc3c78e-177f-4fcd-ab8d-c333497848aa","Type":"ContainerStarted","Data":"0c82cbd59c7a8a28ebfb496e6d839147be4c6469370c138ab28dcca5383f2e97"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.614727 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" event={"ID":"6401ae31-5292-4d4f-8d39-a5682c0d38ee","Type":"ContainerStarted","Data":"782e6781068463fec64c2cd9becc45c6920a17e5a402634e19e412a1240e73cd"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.616109 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.632475 4814 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-tzdzl container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.632549 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" podUID="6401ae31-5292-4d4f-8d39-a5682c0d38ee" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.643372 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj" event={"ID":"5768bed9-9187-4bef-ae5e-bc003793d741","Type":"ContainerStarted","Data":"98de610c83d082d49b993641f59b8c474b1ae7114ac8d99c10625422c6739e01"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.644337 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.645803 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-sbf5t" podStartSLOduration=7.645784982 podStartE2EDuration="7.645784982s" podCreationTimestamp="2026-02-27 16:26:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:56.644192128 +0000 UTC m=+229.096816958" watchObservedRunningTime="2026-02-27 16:26:56.645784982 +0000 UTC m=+229.098409802" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.673605 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twcnw" event={"ID":"634e6fbd-87f9-4205-b01d-90be6eb4a015","Type":"ContainerStarted","Data":"44eb36b48896cf4dc8283fa4629b26c5df510402f1f9ae1e4a21521aa0b96ab9"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.673666 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twcnw" event={"ID":"634e6fbd-87f9-4205-b01d-90be6eb4a015","Type":"ContainerStarted","Data":"696c033a2cd6d9ee72ba2bfd88210bf2430fd7dc78db8609ccb4d7d7bf3ac666"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.674324 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:56 crc kubenswrapper[4814]: E0227 16:26:56.676594 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:57.176579966 +0000 UTC m=+229.629204796 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.694556 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2" event={"ID":"3b269e43-ebd6-4f71-8963-669c60763efa","Type":"ContainerStarted","Data":"39094776090ab68a698f2ab467814e37b383ef38b877d15140894088dc6c815e"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.696025 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gnk2x" podStartSLOduration=163.696015547 podStartE2EDuration="2m43.696015547s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:56.693668796 +0000 UTC m=+229.146293626" watchObservedRunningTime="2026-02-27 16:26:56.696015547 +0000 UTC m=+229.148640377" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.717052 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-6hpl2" event={"ID":"55b73477-5f70-4dcc-a9aa-3e22ac4efb54","Type":"ContainerStarted","Data":"9eb02bd08abb07f47fdd9f930b4265a2b572a757a7ec39e819b0ff94294ab0d1"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.719905 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" podUID="1000b998-ffa1-40c5-8563-0aeb70d172a9" containerName="controller-manager" containerID="cri-o://46c85e5d63ee25a1bc79a3025637a1a712251b11aea6dbf1b31bd48ca586c989" gracePeriod=30 Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.720100 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" event={"ID":"057f89ed-0870-4017-8a0c-a4808d3fb86b","Type":"ContainerStarted","Data":"ac06d0380970a58c7bba32aa62e2af77b11303529121b98296806c6ac027c6c9"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.720130 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.720141 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" event={"ID":"057f89ed-0870-4017-8a0c-a4808d3fb86b","Type":"ContainerStarted","Data":"e155b3559d76e0270e6396280e232852a39143ed28e9d3c2c80f5f89ae9ab8ef"} Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.720215 4814 patch_prober.go:28] interesting pod/downloads-7954f5f757-2n4tf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.720234 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2n4tf" podUID="98112972-5121-4d93-8490-5156a1a2f3ca" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.720376 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" podUID="2b8fa45e-2136-44a1-9a50-f936b5c22417" containerName="route-controller-manager" containerID="cri-o://d16902234ca75a2d7614f859b7ae3631debf752938128ebe79c1da001ff140ff" gracePeriod=30 Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.720647 4814 patch_prober.go:28] interesting pod/console-operator-58897d9998-5xv2z container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.720671 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-5xv2z" podUID="db53a48c-000d-4e92-929e-6c6bf561119a" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.748442 4814 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8tt9s container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:5443/healthz\": dial tcp 10.217.0.43:5443: connect: connection refused" start-of-body= Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.748498 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" podUID="057f89ed-0870-4017-8a0c-a4808d3fb86b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.43:5443/healthz\": dial tcp 10.217.0.43:5443: connect: connection refused" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.775646 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" podStartSLOduration=162.775616196 podStartE2EDuration="2m42.775616196s" podCreationTimestamp="2026-02-27 16:24:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:56.742475981 +0000 UTC m=+229.195100811" watchObservedRunningTime="2026-02-27 16:26:56.775616196 +0000 UTC m=+229.228241026" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.775732 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:26:56 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:26:56 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:26:56 crc kubenswrapper[4814]: healthz check failed Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.775780 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.776780 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:56 crc kubenswrapper[4814]: E0227 16:26:56.776870 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:57.276856288 +0000 UTC m=+229.729481118 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.777180 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.776779 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-ncx8v" podStartSLOduration=163.776774015 podStartE2EDuration="2m43.776774015s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:56.775973448 +0000 UTC m=+229.228598278" watchObservedRunningTime="2026-02-27 16:26:56.776774015 +0000 UTC m=+229.229398845" Feb 27 16:26:56 crc kubenswrapper[4814]: E0227 16:26:56.783347 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:57.283338442 +0000 UTC m=+229.735963272 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.856276 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhq9n" podStartSLOduration=163.856048293 podStartE2EDuration="2m43.856048293s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:56.853953301 +0000 UTC m=+229.306578121" watchObservedRunningTime="2026-02-27 16:26:56.856048293 +0000 UTC m=+229.308673123" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.867190 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.867775 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.879619 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:56 crc kubenswrapper[4814]: E0227 16:26:56.880358 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:57.380341371 +0000 UTC m=+229.832966201 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.981366 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:56 crc kubenswrapper[4814]: E0227 16:26:56.981717 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:57.481700662 +0000 UTC m=+229.934325492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.988830 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj" podStartSLOduration=163.988818467 podStartE2EDuration="2m43.988818467s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:56.986638432 +0000 UTC m=+229.439263262" watchObservedRunningTime="2026-02-27 16:26:56.988818467 +0000 UTC m=+229.441443297" Feb 27 16:26:56 crc kubenswrapper[4814]: I0227 16:26:56.989192 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ws5nw" podStartSLOduration=163.98918725 podStartE2EDuration="2m43.98918725s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:56.91242224 +0000 UTC m=+229.365047070" watchObservedRunningTime="2026-02-27 16:26:56.98918725 +0000 UTC m=+229.441812080" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.051362 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" podStartSLOduration=164.051345717 podStartE2EDuration="2m44.051345717s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:57.034624059 +0000 UTC m=+229.487248889" watchObservedRunningTime="2026-02-27 16:26:57.051345717 +0000 UTC m=+229.503970547" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.077307 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lmzk2" podStartSLOduration=164.077288643 podStartE2EDuration="2m44.077288643s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:57.07606696 +0000 UTC m=+229.528691790" watchObservedRunningTime="2026-02-27 16:26:57.077288643 +0000 UTC m=+229.529913483" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.085799 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:57 crc kubenswrapper[4814]: E0227 16:26:57.086178 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:57.586163549 +0000 UTC m=+230.038788379 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.188907 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:57 crc kubenswrapper[4814]: E0227 16:26:57.189543 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:57.689517948 +0000 UTC m=+230.142142768 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.201793 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-g4vb2" podStartSLOduration=164.201777971 podStartE2EDuration="2m44.201777971s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:57.132770028 +0000 UTC m=+229.585394858" watchObservedRunningTime="2026-02-27 16:26:57.201777971 +0000 UTC m=+229.654402801" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.244927 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twcnw" podStartSLOduration=164.244910621 podStartE2EDuration="2m44.244910621s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:57.243604895 +0000 UTC m=+229.696229725" watchObservedRunningTime="2026-02-27 16:26:57.244910621 +0000 UTC m=+229.697535441" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.245195 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" podStartSLOduration=164.24519072 podStartE2EDuration="2m44.24519072s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:57.204610809 +0000 UTC m=+229.657235639" watchObservedRunningTime="2026-02-27 16:26:57.24519072 +0000 UTC m=+229.697815550" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.268123 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-6hpl2" podStartSLOduration=8.268107012 podStartE2EDuration="8.268107012s" podCreationTimestamp="2026-02-27 16:26:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:57.267663677 +0000 UTC m=+229.720288507" watchObservedRunningTime="2026-02-27 16:26:57.268107012 +0000 UTC m=+229.720731842" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.290968 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:57 crc kubenswrapper[4814]: E0227 16:26:57.291287 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:57.791271951 +0000 UTC m=+230.243896781 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.395081 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:57 crc kubenswrapper[4814]: E0227 16:26:57.395415 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:57.895402067 +0000 UTC m=+230.348026897 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.403283 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.499151 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:57 crc kubenswrapper[4814]: E0227 16:26:57.499438 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:57.999404049 +0000 UTC m=+230.452028889 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.499864 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:57 crc kubenswrapper[4814]: E0227 16:26:57.500292 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:58.000271589 +0000 UTC m=+230.452896419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.517717 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.600719 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.600775 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b8fa45e-2136-44a1-9a50-f936b5c22417-client-ca\") pod \"2b8fa45e-2136-44a1-9a50-f936b5c22417\" (UID: \"2b8fa45e-2136-44a1-9a50-f936b5c22417\") " Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.600856 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4c9r4\" (UniqueName: \"kubernetes.io/projected/2b8fa45e-2136-44a1-9a50-f936b5c22417-kube-api-access-4c9r4\") pod \"2b8fa45e-2136-44a1-9a50-f936b5c22417\" (UID: \"2b8fa45e-2136-44a1-9a50-f936b5c22417\") " Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.600890 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b8fa45e-2136-44a1-9a50-f936b5c22417-serving-cert\") pod \"2b8fa45e-2136-44a1-9a50-f936b5c22417\" (UID: \"2b8fa45e-2136-44a1-9a50-f936b5c22417\") " Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.600931 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b8fa45e-2136-44a1-9a50-f936b5c22417-config\") pod \"2b8fa45e-2136-44a1-9a50-f936b5c22417\" (UID: \"2b8fa45e-2136-44a1-9a50-f936b5c22417\") " Feb 27 16:26:57 crc kubenswrapper[4814]: E0227 16:26:57.601794 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:58.101771174 +0000 UTC m=+230.554396004 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.602356 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b8fa45e-2136-44a1-9a50-f936b5c22417-config" (OuterVolumeSpecName: "config") pod "2b8fa45e-2136-44a1-9a50-f936b5c22417" (UID: "2b8fa45e-2136-44a1-9a50-f936b5c22417"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.603058 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b8fa45e-2136-44a1-9a50-f936b5c22417-client-ca" (OuterVolumeSpecName: "client-ca") pod "2b8fa45e-2136-44a1-9a50-f936b5c22417" (UID: "2b8fa45e-2136-44a1-9a50-f936b5c22417"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.624765 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b8fa45e-2136-44a1-9a50-f936b5c22417-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2b8fa45e-2136-44a1-9a50-f936b5c22417" (UID: "2b8fa45e-2136-44a1-9a50-f936b5c22417"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.633395 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b8fa45e-2136-44a1-9a50-f936b5c22417-kube-api-access-4c9r4" (OuterVolumeSpecName: "kube-api-access-4c9r4") pod "2b8fa45e-2136-44a1-9a50-f936b5c22417" (UID: "2b8fa45e-2136-44a1-9a50-f936b5c22417"). InnerVolumeSpecName "kube-api-access-4c9r4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.650838 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.703108 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.703269 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4c9r4\" (UniqueName: \"kubernetes.io/projected/2b8fa45e-2136-44a1-9a50-f936b5c22417-kube-api-access-4c9r4\") on node \"crc\" DevicePath \"\"" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.703282 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b8fa45e-2136-44a1-9a50-f936b5c22417-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.703291 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b8fa45e-2136-44a1-9a50-f936b5c22417-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.703301 4814 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b8fa45e-2136-44a1-9a50-f936b5c22417-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:26:57 crc kubenswrapper[4814]: E0227 16:26:57.703565 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:58.203553078 +0000 UTC m=+230.656177908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.747395 4814 generic.go:334] "Generic (PLEG): container finished" podID="2b8fa45e-2136-44a1-9a50-f936b5c22417" containerID="d16902234ca75a2d7614f859b7ae3631debf752938128ebe79c1da001ff140ff" exitCode=0 Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.747517 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.747651 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" event={"ID":"2b8fa45e-2136-44a1-9a50-f936b5c22417","Type":"ContainerDied","Data":"d16902234ca75a2d7614f859b7ae3631debf752938128ebe79c1da001ff140ff"} Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.747705 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn" event={"ID":"2b8fa45e-2136-44a1-9a50-f936b5c22417","Type":"ContainerDied","Data":"33622f4e01cdbc4268cb0150db92d2acd38068a099177b9f2ff9cc2d5552393d"} Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.747734 4814 scope.go:117] "RemoveContainer" containerID="d16902234ca75a2d7614f859b7ae3631debf752938128ebe79c1da001ff140ff" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.764847 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" event={"ID":"0f52f92d-a4bd-41b3-88dd-579a355f6e30","Type":"ContainerStarted","Data":"00222a82dfcf79050e0b6c931aed9bcdd65b715a27652ad87702f3838354df05"} Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.766275 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:26:57 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:26:57 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:26:57 crc kubenswrapper[4814]: healthz check failed Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.766313 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.774752 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ncx8v" event={"ID":"ba54dbcb-efbd-467b-aa4e-994fde5e1c14","Type":"ContainerStarted","Data":"a94ba2d6bd5dd00d5078d0a454e4c0d9199e2fc71ff189e9ba5ecd3bda40c86e"} Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.810585 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn"] Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.810638 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gmfbn"] Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.810915 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2qkr\" (UniqueName: \"kubernetes.io/projected/1000b998-ffa1-40c5-8563-0aeb70d172a9-kube-api-access-c2qkr\") pod \"1000b998-ffa1-40c5-8563-0aeb70d172a9\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.810942 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-config\") pod \"1000b998-ffa1-40c5-8563-0aeb70d172a9\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.810999 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1000b998-ffa1-40c5-8563-0aeb70d172a9-serving-cert\") pod \"1000b998-ffa1-40c5-8563-0aeb70d172a9\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.811042 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-client-ca\") pod \"1000b998-ffa1-40c5-8563-0aeb70d172a9\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.811073 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-proxy-ca-bundles\") pod \"1000b998-ffa1-40c5-8563-0aeb70d172a9\" (UID: \"1000b998-ffa1-40c5-8563-0aeb70d172a9\") " Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.812606 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-config" (OuterVolumeSpecName: "config") pod "1000b998-ffa1-40c5-8563-0aeb70d172a9" (UID: "1000b998-ffa1-40c5-8563-0aeb70d172a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.820158 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1000b998-ffa1-40c5-8563-0aeb70d172a9-kube-api-access-c2qkr" (OuterVolumeSpecName: "kube-api-access-c2qkr") pod "1000b998-ffa1-40c5-8563-0aeb70d172a9" (UID: "1000b998-ffa1-40c5-8563-0aeb70d172a9"). InnerVolumeSpecName "kube-api-access-c2qkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.824842 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1000b998-ffa1-40c5-8563-0aeb70d172a9" (UID: "1000b998-ffa1-40c5-8563-0aeb70d172a9"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.825096 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-client-ca" (OuterVolumeSpecName: "client-ca") pod "1000b998-ffa1-40c5-8563-0aeb70d172a9" (UID: "1000b998-ffa1-40c5-8563-0aeb70d172a9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:26:57 crc kubenswrapper[4814]: E0227 16:26:57.827194 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:58.327164907 +0000 UTC m=+230.779789737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.825571 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pwdrv" event={"ID":"fc590a1b-ba51-421c-a089-fb0def054fb2","Type":"ContainerStarted","Data":"00d0455a959836d5400b460da6819fb6e9d8ecbf013f270727f099faebae0e90"} Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.825987 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.827942 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.828012 4814 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.828023 4814 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.828032 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2qkr\" (UniqueName: \"kubernetes.io/projected/1000b998-ffa1-40c5-8563-0aeb70d172a9-kube-api-access-c2qkr\") on node \"crc\" DevicePath \"\"" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.828041 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1000b998-ffa1-40c5-8563-0aeb70d172a9-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:26:57 crc kubenswrapper[4814]: E0227 16:26:57.828304 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:58.328294386 +0000 UTC m=+230.780919216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.827559 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1000b998-ffa1-40c5-8563-0aeb70d172a9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1000b998-ffa1-40c5-8563-0aeb70d172a9" (UID: "1000b998-ffa1-40c5-8563-0aeb70d172a9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.845417 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qnsbd" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.845784 4814 scope.go:117] "RemoveContainer" containerID="d16902234ca75a2d7614f859b7ae3631debf752938128ebe79c1da001ff140ff" Feb 27 16:26:57 crc kubenswrapper[4814]: E0227 16:26:57.852643 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d16902234ca75a2d7614f859b7ae3631debf752938128ebe79c1da001ff140ff\": container with ID starting with d16902234ca75a2d7614f859b7ae3631debf752938128ebe79c1da001ff140ff not found: ID does not exist" containerID="d16902234ca75a2d7614f859b7ae3631debf752938128ebe79c1da001ff140ff" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.852703 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d16902234ca75a2d7614f859b7ae3631debf752938128ebe79c1da001ff140ff"} err="failed to get container status \"d16902234ca75a2d7614f859b7ae3631debf752938128ebe79c1da001ff140ff\": rpc error: code = NotFound desc = could not find container \"d16902234ca75a2d7614f859b7ae3631debf752938128ebe79c1da001ff140ff\": container with ID starting with d16902234ca75a2d7614f859b7ae3631debf752938128ebe79c1da001ff140ff not found: ID does not exist" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.862968 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-pwdrv" podStartSLOduration=164.862937412 podStartE2EDuration="2m44.862937412s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:57.862798917 +0000 UTC m=+230.315423737" watchObservedRunningTime="2026-02-27 16:26:57.862937412 +0000 UTC m=+230.315562242" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.878342 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx" event={"ID":"d32d7bef-558f-4dfd-9260-408cb47fc461","Type":"ContainerStarted","Data":"8da7cb9e070a6141913fab7b81fbe07ea866924f8fd58ceb6429c4ecdd8d6ed3"} Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.899481 4814 generic.go:334] "Generic (PLEG): container finished" podID="1000b998-ffa1-40c5-8563-0aeb70d172a9" containerID="46c85e5d63ee25a1bc79a3025637a1a712251b11aea6dbf1b31bd48ca586c989" exitCode=0 Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.899598 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.899705 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" event={"ID":"1000b998-ffa1-40c5-8563-0aeb70d172a9","Type":"ContainerDied","Data":"46c85e5d63ee25a1bc79a3025637a1a712251b11aea6dbf1b31bd48ca586c989"} Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.899750 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4m8vp" event={"ID":"1000b998-ffa1-40c5-8563-0aeb70d172a9","Type":"ContainerDied","Data":"1deaf117186b7d1efa4ddcd141c77174ac58f9bca95b24a5d9fad1354e272549"} Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.899769 4814 scope.go:117] "RemoveContainer" containerID="46c85e5d63ee25a1bc79a3025637a1a712251b11aea6dbf1b31bd48ca586c989" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.930113 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rs9mx" podStartSLOduration=164.930095601 podStartE2EDuration="2m44.930095601s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:57.929907274 +0000 UTC m=+230.382532104" watchObservedRunningTime="2026-02-27 16:26:57.930095601 +0000 UTC m=+230.382720431" Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.930943 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.931406 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1000b998-ffa1-40c5-8563-0aeb70d172a9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:26:57 crc kubenswrapper[4814]: E0227 16:26:57.932497 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:58.432477373 +0000 UTC m=+230.885102203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.971539 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" event={"ID":"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b","Type":"ContainerStarted","Data":"bf771dc3a3519a84faa9e99f777a356e49a7dda0413fd5cee86ac5844bc5a872"} Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.971574 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" event={"ID":"e54ad2c6-9634-4d0e-a38d-6a2d0cc04a6b","Type":"ContainerStarted","Data":"457e661595e61b10b72755346c59d7a6163adef650ffe984a8fef4d8e3814bd8"} Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.972349 4814 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mjlld container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Feb 27 16:26:57 crc kubenswrapper[4814]: I0227 16:26:57.972422 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" podUID="dfa49cf2-dc52-4c27-8486-ec1efec61744" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.001314 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tzdzl" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.003713 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hl9wj" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.004711 4814 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8tt9s container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:5443/healthz\": dial tcp 10.217.0.43:5443: connect: connection refused" start-of-body= Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.004789 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" podUID="057f89ed-0870-4017-8a0c-a4808d3fb86b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.43:5443/healthz\": dial tcp 10.217.0.43:5443: connect: connection refused" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.007757 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-46nl6" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.018810 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4m8vp"] Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.033810 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:58 crc kubenswrapper[4814]: E0227 16:26:58.035314 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:58.535288224 +0000 UTC m=+230.987913054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.041582 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4m8vp"] Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.078529 4814 scope.go:117] "RemoveContainer" containerID="46c85e5d63ee25a1bc79a3025637a1a712251b11aea6dbf1b31bd48ca586c989" Feb 27 16:26:58 crc kubenswrapper[4814]: E0227 16:26:58.086412 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46c85e5d63ee25a1bc79a3025637a1a712251b11aea6dbf1b31bd48ca586c989\": container with ID starting with 46c85e5d63ee25a1bc79a3025637a1a712251b11aea6dbf1b31bd48ca586c989 not found: ID does not exist" containerID="46c85e5d63ee25a1bc79a3025637a1a712251b11aea6dbf1b31bd48ca586c989" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.086487 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46c85e5d63ee25a1bc79a3025637a1a712251b11aea6dbf1b31bd48ca586c989"} err="failed to get container status \"46c85e5d63ee25a1bc79a3025637a1a712251b11aea6dbf1b31bd48ca586c989\": rpc error: code = NotFound desc = could not find container \"46c85e5d63ee25a1bc79a3025637a1a712251b11aea6dbf1b31bd48ca586c989\": container with ID starting with 46c85e5d63ee25a1bc79a3025637a1a712251b11aea6dbf1b31bd48ca586c989 not found: ID does not exist" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.109200 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" podStartSLOduration=165.109182545 podStartE2EDuration="2m45.109182545s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:26:58.107790557 +0000 UTC m=+230.560415387" watchObservedRunningTime="2026-02-27 16:26:58.109182545 +0000 UTC m=+230.561807375" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.136018 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:58 crc kubenswrapper[4814]: E0227 16:26:58.138028 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:58.63800551 +0000 UTC m=+231.090630340 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.240992 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:58 crc kubenswrapper[4814]: E0227 16:26:58.241317 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:58.741304567 +0000 UTC m=+231.193929397 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.341747 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:58 crc kubenswrapper[4814]: E0227 16:26:58.342461 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:58.8424458 +0000 UTC m=+231.295070630 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.444826 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:58 crc kubenswrapper[4814]: E0227 16:26:58.445111 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:58.945099415 +0000 UTC m=+231.397724245 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.469291 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-64967bf4f7-rmz25"] Feb 27 16:26:58 crc kubenswrapper[4814]: E0227 16:26:58.469471 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b8fa45e-2136-44a1-9a50-f936b5c22417" containerName="route-controller-manager" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.469498 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b8fa45e-2136-44a1-9a50-f936b5c22417" containerName="route-controller-manager" Feb 27 16:26:58 crc kubenswrapper[4814]: E0227 16:26:58.469511 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1000b998-ffa1-40c5-8563-0aeb70d172a9" containerName="controller-manager" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.469519 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="1000b998-ffa1-40c5-8563-0aeb70d172a9" containerName="controller-manager" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.469612 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="1000b998-ffa1-40c5-8563-0aeb70d172a9" containerName="controller-manager" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.469628 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b8fa45e-2136-44a1-9a50-f936b5c22417" containerName="route-controller-manager" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.469926 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.470150 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm"] Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.470566 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.479231 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.479516 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.479627 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.479655 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.479913 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.479882 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.480221 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.480342 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.480469 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.480585 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.480824 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.480913 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.494382 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.515593 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1000b998-ffa1-40c5-8563-0aeb70d172a9" path="/var/lib/kubelet/pods/1000b998-ffa1-40c5-8563-0aeb70d172a9/volumes" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.516314 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b8fa45e-2136-44a1-9a50-f936b5c22417" path="/var/lib/kubelet/pods/2b8fa45e-2136-44a1-9a50-f936b5c22417/volumes" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.516747 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-64967bf4f7-rmz25"] Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.516778 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm"] Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.545698 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:58 crc kubenswrapper[4814]: E0227 16:26:58.546103 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:59.046088522 +0000 UTC m=+231.498713352 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.646886 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r68sl\" (UniqueName: \"kubernetes.io/projected/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-kube-api-access-r68sl\") pod \"route-controller-manager-687b4d5955-4wznm\" (UID: \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\") " pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.646926 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-serving-cert\") pod \"controller-manager-64967bf4f7-rmz25\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.646947 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-config\") pod \"controller-manager-64967bf4f7-rmz25\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.646965 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-client-ca\") pod \"controller-manager-64967bf4f7-rmz25\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.646985 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-proxy-ca-bundles\") pod \"controller-manager-64967bf4f7-rmz25\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.647017 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-client-ca\") pod \"route-controller-manager-687b4d5955-4wznm\" (UID: \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\") " pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.647035 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-config\") pod \"route-controller-manager-687b4d5955-4wznm\" (UID: \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\") " pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.647056 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-serving-cert\") pod \"route-controller-manager-687b4d5955-4wznm\" (UID: \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\") " pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.647076 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.647095 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrqsj\" (UniqueName: \"kubernetes.io/projected/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-kube-api-access-hrqsj\") pod \"controller-manager-64967bf4f7-rmz25\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:58 crc kubenswrapper[4814]: E0227 16:26:58.647448 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:59.147435512 +0000 UTC m=+231.600060332 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.748340 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.748538 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-client-ca\") pod \"route-controller-manager-687b4d5955-4wznm\" (UID: \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\") " pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.748564 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-config\") pod \"route-controller-manager-687b4d5955-4wznm\" (UID: \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\") " pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.748588 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-serving-cert\") pod \"route-controller-manager-687b4d5955-4wznm\" (UID: \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\") " pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.748615 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrqsj\" (UniqueName: \"kubernetes.io/projected/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-kube-api-access-hrqsj\") pod \"controller-manager-64967bf4f7-rmz25\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.748686 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r68sl\" (UniqueName: \"kubernetes.io/projected/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-kube-api-access-r68sl\") pod \"route-controller-manager-687b4d5955-4wznm\" (UID: \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\") " pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.748705 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-serving-cert\") pod \"controller-manager-64967bf4f7-rmz25\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.748738 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-config\") pod \"controller-manager-64967bf4f7-rmz25\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.748766 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-client-ca\") pod \"controller-manager-64967bf4f7-rmz25\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.748799 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-proxy-ca-bundles\") pod \"controller-manager-64967bf4f7-rmz25\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.749990 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-proxy-ca-bundles\") pod \"controller-manager-64967bf4f7-rmz25\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.750289 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-client-ca\") pod \"route-controller-manager-687b4d5955-4wznm\" (UID: \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\") " pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:26:58 crc kubenswrapper[4814]: E0227 16:26:58.750397 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:59.250374666 +0000 UTC m=+231.702999486 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.750610 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-config\") pod \"route-controller-manager-687b4d5955-4wznm\" (UID: \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\") " pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.751049 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-client-ca\") pod \"controller-manager-64967bf4f7-rmz25\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.751384 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-config\") pod \"controller-manager-64967bf4f7-rmz25\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.757407 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-serving-cert\") pod \"controller-manager-64967bf4f7-rmz25\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.771463 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:26:58 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:26:58 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:26:58 crc kubenswrapper[4814]: healthz check failed Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.771508 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.784840 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r68sl\" (UniqueName: \"kubernetes.io/projected/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-kube-api-access-r68sl\") pod \"route-controller-manager-687b4d5955-4wznm\" (UID: \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\") " pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.791976 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-serving-cert\") pod \"route-controller-manager-687b4d5955-4wznm\" (UID: \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\") " pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.802983 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.804070 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrqsj\" (UniqueName: \"kubernetes.io/projected/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-kube-api-access-hrqsj\") pod \"controller-manager-64967bf4f7-rmz25\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.835628 4814 ???:1] "http: TLS handshake error from 192.168.126.11:56336: no serving certificate available for the kubelet" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.849844 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:58 crc kubenswrapper[4814]: E0227 16:26:58.850202 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:59.350176173 +0000 UTC m=+231.802801003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.902662 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cm8p5"] Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.903887 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.906724 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.951237 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:58 crc kubenswrapper[4814]: E0227 16:26:58.951586 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:59.451571384 +0000 UTC m=+231.904196214 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:58 crc kubenswrapper[4814]: I0227 16:26:58.977538 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cm8p5"] Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.049792 4814 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mjlld container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.049836 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" podUID="dfa49cf2-dc52-4c27-8486-ec1efec61744" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.052158 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-utilities\") pod \"certified-operators-cm8p5\" (UID: \"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375\") " pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.052202 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-catalog-content\") pod \"certified-operators-cm8p5\" (UID: \"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375\") " pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.052234 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.052266 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5cld\" (UniqueName: \"kubernetes.io/projected/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-kube-api-access-h5cld\") pod \"certified-operators-cm8p5\" (UID: \"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375\") " pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:26:59 crc kubenswrapper[4814]: E0227 16:26:59.052602 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:59.552587402 +0000 UTC m=+232.005212232 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.078549 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sk6rs"] Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.090159 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.099720 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.100649 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.107217 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sk6rs"] Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.152829 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.153193 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-utilities\") pod \"certified-operators-cm8p5\" (UID: \"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375\") " pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.153355 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-catalog-content\") pod \"certified-operators-cm8p5\" (UID: \"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375\") " pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.153432 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5cld\" (UniqueName: \"kubernetes.io/projected/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-kube-api-access-h5cld\") pod \"certified-operators-cm8p5\" (UID: \"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375\") " pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:26:59 crc kubenswrapper[4814]: E0227 16:26:59.154938 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:59.654894845 +0000 UTC m=+232.107519665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.158428 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-catalog-content\") pod \"certified-operators-cm8p5\" (UID: \"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375\") " pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.158638 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-utilities\") pod \"certified-operators-cm8p5\" (UID: \"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375\") " pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.223563 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5cld\" (UniqueName: \"kubernetes.io/projected/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-kube-api-access-h5cld\") pod \"certified-operators-cm8p5\" (UID: \"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375\") " pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.236509 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm"] Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.242529 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.255397 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41b851a0-9bda-4e20-a692-5f796f55a749-catalog-content\") pod \"community-operators-sk6rs\" (UID: \"41b851a0-9bda-4e20-a692-5f796f55a749\") " pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.255447 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgd64\" (UniqueName: \"kubernetes.io/projected/41b851a0-9bda-4e20-a692-5f796f55a749-kube-api-access-vgd64\") pod \"community-operators-sk6rs\" (UID: \"41b851a0-9bda-4e20-a692-5f796f55a749\") " pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.255491 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41b851a0-9bda-4e20-a692-5f796f55a749-utilities\") pod \"community-operators-sk6rs\" (UID: \"41b851a0-9bda-4e20-a692-5f796f55a749\") " pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.255567 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:59 crc kubenswrapper[4814]: E0227 16:26:59.255967 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:59.755951425 +0000 UTC m=+232.208576255 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.280163 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c5wns"] Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.281160 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.313755 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c5wns"] Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.356889 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.357275 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rddgl\" (UniqueName: \"kubernetes.io/projected/d3a647cf-eb96-40b8-982c-8e8b869e57c4-kube-api-access-rddgl\") pod \"certified-operators-c5wns\" (UID: \"d3a647cf-eb96-40b8-982c-8e8b869e57c4\") " pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.357326 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41b851a0-9bda-4e20-a692-5f796f55a749-catalog-content\") pod \"community-operators-sk6rs\" (UID: \"41b851a0-9bda-4e20-a692-5f796f55a749\") " pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.357350 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a647cf-eb96-40b8-982c-8e8b869e57c4-catalog-content\") pod \"certified-operators-c5wns\" (UID: \"d3a647cf-eb96-40b8-982c-8e8b869e57c4\") " pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.357370 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgd64\" (UniqueName: \"kubernetes.io/projected/41b851a0-9bda-4e20-a692-5f796f55a749-kube-api-access-vgd64\") pod \"community-operators-sk6rs\" (UID: \"41b851a0-9bda-4e20-a692-5f796f55a749\") " pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.357406 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41b851a0-9bda-4e20-a692-5f796f55a749-utilities\") pod \"community-operators-sk6rs\" (UID: \"41b851a0-9bda-4e20-a692-5f796f55a749\") " pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.357424 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a647cf-eb96-40b8-982c-8e8b869e57c4-utilities\") pod \"certified-operators-c5wns\" (UID: \"d3a647cf-eb96-40b8-982c-8e8b869e57c4\") " pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:26:59 crc kubenswrapper[4814]: E0227 16:26:59.357637 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:26:59.857612185 +0000 UTC m=+232.310237015 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.358627 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41b851a0-9bda-4e20-a692-5f796f55a749-catalog-content\") pod \"community-operators-sk6rs\" (UID: \"41b851a0-9bda-4e20-a692-5f796f55a749\") " pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.360103 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41b851a0-9bda-4e20-a692-5f796f55a749-utilities\") pod \"community-operators-sk6rs\" (UID: \"41b851a0-9bda-4e20-a692-5f796f55a749\") " pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.378819 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgd64\" (UniqueName: \"kubernetes.io/projected/41b851a0-9bda-4e20-a692-5f796f55a749-kube-api-access-vgd64\") pod \"community-operators-sk6rs\" (UID: \"41b851a0-9bda-4e20-a692-5f796f55a749\") " pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.432147 4814 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.432325 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.458516 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.458575 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rddgl\" (UniqueName: \"kubernetes.io/projected/d3a647cf-eb96-40b8-982c-8e8b869e57c4-kube-api-access-rddgl\") pod \"certified-operators-c5wns\" (UID: \"d3a647cf-eb96-40b8-982c-8e8b869e57c4\") " pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.458636 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a647cf-eb96-40b8-982c-8e8b869e57c4-catalog-content\") pod \"certified-operators-c5wns\" (UID: \"d3a647cf-eb96-40b8-982c-8e8b869e57c4\") " pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.458679 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a647cf-eb96-40b8-982c-8e8b869e57c4-utilities\") pod \"certified-operators-c5wns\" (UID: \"d3a647cf-eb96-40b8-982c-8e8b869e57c4\") " pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:26:59 crc kubenswrapper[4814]: E0227 16:26:59.459534 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:26:59.959519464 +0000 UTC m=+232.412144294 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.460421 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a647cf-eb96-40b8-982c-8e8b869e57c4-catalog-content\") pod \"certified-operators-c5wns\" (UID: \"d3a647cf-eb96-40b8-982c-8e8b869e57c4\") " pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.464415 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a647cf-eb96-40b8-982c-8e8b869e57c4-utilities\") pod \"certified-operators-c5wns\" (UID: \"d3a647cf-eb96-40b8-982c-8e8b869e57c4\") " pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.484241 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-972lm"] Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.485608 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-972lm" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.490170 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rddgl\" (UniqueName: \"kubernetes.io/projected/d3a647cf-eb96-40b8-982c-8e8b869e57c4-kube-api-access-rddgl\") pod \"certified-operators-c5wns\" (UID: \"d3a647cf-eb96-40b8-982c-8e8b869e57c4\") " pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.493560 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-972lm"] Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.560177 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.560410 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cdfdb15-014d-40fc-9a7b-a905278f1249-catalog-content\") pod \"community-operators-972lm\" (UID: \"2cdfdb15-014d-40fc-9a7b-a905278f1249\") " pod="openshift-marketplace/community-operators-972lm" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.560487 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cdfdb15-014d-40fc-9a7b-a905278f1249-utilities\") pod \"community-operators-972lm\" (UID: \"2cdfdb15-014d-40fc-9a7b-a905278f1249\") " pod="openshift-marketplace/community-operators-972lm" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.560508 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmsmv\" (UniqueName: \"kubernetes.io/projected/2cdfdb15-014d-40fc-9a7b-a905278f1249-kube-api-access-pmsmv\") pod \"community-operators-972lm\" (UID: \"2cdfdb15-014d-40fc-9a7b-a905278f1249\") " pod="openshift-marketplace/community-operators-972lm" Feb 27 16:26:59 crc kubenswrapper[4814]: E0227 16:26:59.560558 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 16:27:00.060537173 +0000 UTC m=+232.513162003 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.598608 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-64967bf4f7-rmz25"] Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.624273 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.663611 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cdfdb15-014d-40fc-9a7b-a905278f1249-catalog-content\") pod \"community-operators-972lm\" (UID: \"2cdfdb15-014d-40fc-9a7b-a905278f1249\") " pod="openshift-marketplace/community-operators-972lm" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.663923 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.663967 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cdfdb15-014d-40fc-9a7b-a905278f1249-utilities\") pod \"community-operators-972lm\" (UID: \"2cdfdb15-014d-40fc-9a7b-a905278f1249\") " pod="openshift-marketplace/community-operators-972lm" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.663986 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmsmv\" (UniqueName: \"kubernetes.io/projected/2cdfdb15-014d-40fc-9a7b-a905278f1249-kube-api-access-pmsmv\") pod \"community-operators-972lm\" (UID: \"2cdfdb15-014d-40fc-9a7b-a905278f1249\") " pod="openshift-marketplace/community-operators-972lm" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.664761 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cdfdb15-014d-40fc-9a7b-a905278f1249-catalog-content\") pod \"community-operators-972lm\" (UID: \"2cdfdb15-014d-40fc-9a7b-a905278f1249\") " pod="openshift-marketplace/community-operators-972lm" Feb 27 16:26:59 crc kubenswrapper[4814]: E0227 16:26:59.665093 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 16:27:00.165070272 +0000 UTC m=+232.617695172 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9ws8" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.665134 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cdfdb15-014d-40fc-9a7b-a905278f1249-utilities\") pod \"community-operators-972lm\" (UID: \"2cdfdb15-014d-40fc-9a7b-a905278f1249\") " pod="openshift-marketplace/community-operators-972lm" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.710374 4814 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-27T16:26:59.43217534Z","Handler":null,"Name":""} Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.712561 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmsmv\" (UniqueName: \"kubernetes.io/projected/2cdfdb15-014d-40fc-9a7b-a905278f1249-kube-api-access-pmsmv\") pod \"community-operators-972lm\" (UID: \"2cdfdb15-014d-40fc-9a7b-a905278f1249\") " pod="openshift-marketplace/community-operators-972lm" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.713529 4814 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.713578 4814 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.728938 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cm8p5"] Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.760609 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:26:59 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:26:59 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:26:59 crc kubenswrapper[4814]: healthz check failed Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.760692 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.766460 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.778053 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sk6rs"] Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.823921 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-972lm" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.925204 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.930966 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c5wns"] Feb 27 16:26:59 crc kubenswrapper[4814]: W0227 16:26:59.937735 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1ed29a3_a2f5_46e1_ac86_b3adc8c52375.slice/crio-6683dda6eb1039822dea692ee60f4fd262751c4125c7f2bad504712dcece7f7d WatchSource:0}: Error finding container 6683dda6eb1039822dea692ee60f4fd262751c4125c7f2bad504712dcece7f7d: Status 404 returned error can't find the container with id 6683dda6eb1039822dea692ee60f4fd262751c4125c7f2bad504712dcece7f7d Feb 27 16:26:59 crc kubenswrapper[4814]: W0227 16:26:59.962803 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41b851a0_9bda_4e20_a692_5f796f55a749.slice/crio-4f549b4cfbcf24cc2938fb684e94cf90a0493916225b398319d177632e456eab WatchSource:0}: Error finding container 4f549b4cfbcf24cc2938fb684e94cf90a0493916225b398319d177632e456eab: Status 404 returned error can't find the container with id 4f549b4cfbcf24cc2938fb684e94cf90a0493916225b398319d177632e456eab Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.970615 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.974025 4814 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 27 16:26:59 crc kubenswrapper[4814]: I0227 16:26:59.974064 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.014597 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9ws8\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.072398 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cm8p5" event={"ID":"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375","Type":"ContainerStarted","Data":"6683dda6eb1039822dea692ee60f4fd262751c4125c7f2bad504712dcece7f7d"} Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.077021 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" event={"ID":"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9","Type":"ContainerStarted","Data":"213dc6cee9532429c439ee6d28414db5c00acf5519c9ef6c258e764bf8efe044"} Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.087790 4814 generic.go:334] "Generic (PLEG): container finished" podID="d87f2169-8a70-42de-aaf5-7728ff95fa50" containerID="b86ccc9ec3b77950f4a33f3517ef3218c0cf88bf4a5d8200fb3cda18db954532" exitCode=0 Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.087914 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" event={"ID":"d87f2169-8a70-42de-aaf5-7728ff95fa50","Type":"ContainerDied","Data":"b86ccc9ec3b77950f4a33f3517ef3218c0cf88bf4a5d8200fb3cda18db954532"} Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.109825 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" event={"ID":"0f52f92d-a4bd-41b3-88dd-579a355f6e30","Type":"ContainerStarted","Data":"50a874fdbcb4bee08b789bb0fea35a1b7f0899c222b44724bdfddb190171974f"} Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.109864 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" event={"ID":"0f52f92d-a4bd-41b3-88dd-579a355f6e30","Type":"ContainerStarted","Data":"4d54ce28ee591ef8c33e61fe4736c68bfaf4f65492c723c8e48b474b13ef5bfa"} Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.117119 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sk6rs" event={"ID":"41b851a0-9bda-4e20-a692-5f796f55a749","Type":"ContainerStarted","Data":"4f549b4cfbcf24cc2938fb684e94cf90a0493916225b398319d177632e456eab"} Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.136598 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c5wns" event={"ID":"d3a647cf-eb96-40b8-982c-8e8b869e57c4","Type":"ContainerStarted","Data":"c18a0d2234e658205a945f939203e622c10efd7c68e81f88e2e5f8143bebbc7e"} Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.159321 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" event={"ID":"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c","Type":"ContainerStarted","Data":"526183fdafe1c67c9473802ba8555164094bdf32a27178105649c4bb6176725b"} Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.159354 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" event={"ID":"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c","Type":"ContainerStarted","Data":"fc1fec091a5ae07f6b635962e96f928452d581369755e5690ea05ff1a08e72e0"} Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.160024 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.198301 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" podStartSLOduration=3.198284765 podStartE2EDuration="3.198284765s" podCreationTimestamp="2026-02-27 16:26:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:27:00.198056507 +0000 UTC m=+232.650681337" watchObservedRunningTime="2026-02-27 16:27:00.198284765 +0000 UTC m=+232.650909595" Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.233172 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.367191 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-972lm"] Feb 27 16:27:00 crc kubenswrapper[4814]: W0227 16:27:00.430711 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cdfdb15_014d_40fc_9a7b_a905278f1249.slice/crio-113a0bc7b7414845c6dfa565ec8e4b4b54dd39570fee9bed75caeecf391e390d WatchSource:0}: Error finding container 113a0bc7b7414845c6dfa565ec8e4b4b54dd39570fee9bed75caeecf391e390d: Status 404 returned error can't find the container with id 113a0bc7b7414845c6dfa565ec8e4b4b54dd39570fee9bed75caeecf391e390d Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.469071 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.507782 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.569646 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9ws8"] Feb 27 16:27:00 crc kubenswrapper[4814]: W0227 16:27:00.602036 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b4a4d02_59a6_4f74_a4cc_2154d1115361.slice/crio-fd04451f58791f8461df970d3d11dff3a454c2366400d4f43cffdb7695978e44 WatchSource:0}: Error finding container fd04451f58791f8461df970d3d11dff3a454c2366400d4f43cffdb7695978e44: Status 404 returned error can't find the container with id fd04451f58791f8461df970d3d11dff3a454c2366400d4f43cffdb7695978e44 Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.760126 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:27:00 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:27:00 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:27:00 crc kubenswrapper[4814]: healthz check failed Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.760707 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.879001 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4zbl6"] Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.880203 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.882918 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.890776 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4zbl6"] Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.994074 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5df9c141-ab4f-40e3-b733-da2a2712a326-utilities\") pod \"redhat-marketplace-4zbl6\" (UID: \"5df9c141-ab4f-40e3-b733-da2a2712a326\") " pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.994139 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5df9c141-ab4f-40e3-b733-da2a2712a326-catalog-content\") pod \"redhat-marketplace-4zbl6\" (UID: \"5df9c141-ab4f-40e3-b733-da2a2712a326\") " pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:27:00 crc kubenswrapper[4814]: I0227 16:27:00.994675 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpb22\" (UniqueName: \"kubernetes.io/projected/5df9c141-ab4f-40e3-b733-da2a2712a326-kube-api-access-tpb22\") pod \"redhat-marketplace-4zbl6\" (UID: \"5df9c141-ab4f-40e3-b733-da2a2712a326\") " pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.095833 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpb22\" (UniqueName: \"kubernetes.io/projected/5df9c141-ab4f-40e3-b733-da2a2712a326-kube-api-access-tpb22\") pod \"redhat-marketplace-4zbl6\" (UID: \"5df9c141-ab4f-40e3-b733-da2a2712a326\") " pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.095943 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5df9c141-ab4f-40e3-b733-da2a2712a326-utilities\") pod \"redhat-marketplace-4zbl6\" (UID: \"5df9c141-ab4f-40e3-b733-da2a2712a326\") " pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.095974 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5df9c141-ab4f-40e3-b733-da2a2712a326-catalog-content\") pod \"redhat-marketplace-4zbl6\" (UID: \"5df9c141-ab4f-40e3-b733-da2a2712a326\") " pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.096500 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5df9c141-ab4f-40e3-b733-da2a2712a326-catalog-content\") pod \"redhat-marketplace-4zbl6\" (UID: \"5df9c141-ab4f-40e3-b733-da2a2712a326\") " pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.096677 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5df9c141-ab4f-40e3-b733-da2a2712a326-utilities\") pod \"redhat-marketplace-4zbl6\" (UID: \"5df9c141-ab4f-40e3-b733-da2a2712a326\") " pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.132664 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpb22\" (UniqueName: \"kubernetes.io/projected/5df9c141-ab4f-40e3-b733-da2a2712a326-kube-api-access-tpb22\") pod \"redhat-marketplace-4zbl6\" (UID: \"5df9c141-ab4f-40e3-b733-da2a2712a326\") " pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.171667 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" event={"ID":"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9","Type":"ContainerStarted","Data":"3ecfe545b52140b6f8d8e2da63d6dd1c7dbf2fe5f9fed179b29c2932932e24e2"} Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.172985 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.182026 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.182505 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" event={"ID":"0f52f92d-a4bd-41b3-88dd-579a355f6e30","Type":"ContainerStarted","Data":"d2dc30b43fbc2dcf1aaa386d22e4b8de7356d8249d6f3ef83fb839c3f58ed78c"} Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.184195 4814 generic.go:334] "Generic (PLEG): container finished" podID="41b851a0-9bda-4e20-a692-5f796f55a749" containerID="c91bbb45267bcaa1b07196607bd447558099939728a7922c29ef04c8ec6303b7" exitCode=0 Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.184243 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sk6rs" event={"ID":"41b851a0-9bda-4e20-a692-5f796f55a749","Type":"ContainerDied","Data":"c91bbb45267bcaa1b07196607bd447558099939728a7922c29ef04c8ec6303b7"} Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.187103 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" event={"ID":"2b4a4d02-59a6-4f74-a4cc-2154d1115361","Type":"ContainerStarted","Data":"e3d9c465ec3deb6bf096badcd6f47ff7aab7c32dd2b06a6b8d5f07b4efac3a8a"} Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.187143 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" event={"ID":"2b4a4d02-59a6-4f74-a4cc-2154d1115361","Type":"ContainerStarted","Data":"fd04451f58791f8461df970d3d11dff3a454c2366400d4f43cffdb7695978e44"} Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.188888 4814 generic.go:334] "Generic (PLEG): container finished" podID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" containerID="3749fb0334366dc48359430c6354be4ed08fcf1840c7a770d9874cfb029e4a0f" exitCode=0 Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.188952 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c5wns" event={"ID":"d3a647cf-eb96-40b8-982c-8e8b869e57c4","Type":"ContainerDied","Data":"3749fb0334366dc48359430c6354be4ed08fcf1840c7a770d9874cfb029e4a0f"} Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.190090 4814 generic.go:334] "Generic (PLEG): container finished" podID="2cdfdb15-014d-40fc-9a7b-a905278f1249" containerID="d49a517eb4c3d0ee712ade3917d8b9c94bcf5a6260f8f0f00ccda868f1d3c25c" exitCode=0 Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.190162 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-972lm" event={"ID":"2cdfdb15-014d-40fc-9a7b-a905278f1249","Type":"ContainerDied","Data":"d49a517eb4c3d0ee712ade3917d8b9c94bcf5a6260f8f0f00ccda868f1d3c25c"} Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.190190 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-972lm" event={"ID":"2cdfdb15-014d-40fc-9a7b-a905278f1249","Type":"ContainerStarted","Data":"113a0bc7b7414845c6dfa565ec8e4b4b54dd39570fee9bed75caeecf391e390d"} Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.192411 4814 generic.go:334] "Generic (PLEG): container finished" podID="f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" containerID="c56f2147f422717aab783873ae75de9d5cdc15359a4cd3f56720360d84a63586" exitCode=0 Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.192503 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cm8p5" event={"ID":"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375","Type":"ContainerDied","Data":"c56f2147f422717aab783873ae75de9d5cdc15359a4cd3f56720360d84a63586"} Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.199100 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.204893 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.214115 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" podStartSLOduration=4.214088801 podStartE2EDuration="4.214088801s" podCreationTimestamp="2026-02-27 16:26:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:27:01.203218746 +0000 UTC m=+233.655843576" watchObservedRunningTime="2026-02-27 16:27:01.214088801 +0000 UTC m=+233.666713631" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.282034 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-h8mkp" podStartSLOduration=12.281985326 podStartE2EDuration="12.281985326s" podCreationTimestamp="2026-02-27 16:26:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:27:01.272594161 +0000 UTC m=+233.725218991" watchObservedRunningTime="2026-02-27 16:27:01.281985326 +0000 UTC m=+233.734610156" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.296904 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ltqgw"] Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.299004 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.309475 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" podStartSLOduration=168.309453404 podStartE2EDuration="2m48.309453404s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:27:01.307983904 +0000 UTC m=+233.760608734" watchObservedRunningTime="2026-02-27 16:27:01.309453404 +0000 UTC m=+233.762078234" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.327497 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ltqgw"] Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.407573 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44a897a0-3e5d-4c66-b159-a8830705e7cf-catalog-content\") pod \"redhat-marketplace-ltqgw\" (UID: \"44a897a0-3e5d-4c66-b159-a8830705e7cf\") " pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.407667 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgs7w\" (UniqueName: \"kubernetes.io/projected/44a897a0-3e5d-4c66-b159-a8830705e7cf-kube-api-access-hgs7w\") pod \"redhat-marketplace-ltqgw\" (UID: \"44a897a0-3e5d-4c66-b159-a8830705e7cf\") " pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.407695 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44a897a0-3e5d-4c66-b159-a8830705e7cf-utilities\") pod \"redhat-marketplace-ltqgw\" (UID: \"44a897a0-3e5d-4c66-b159-a8830705e7cf\") " pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.510118 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgs7w\" (UniqueName: \"kubernetes.io/projected/44a897a0-3e5d-4c66-b159-a8830705e7cf-kube-api-access-hgs7w\") pod \"redhat-marketplace-ltqgw\" (UID: \"44a897a0-3e5d-4c66-b159-a8830705e7cf\") " pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.510172 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44a897a0-3e5d-4c66-b159-a8830705e7cf-utilities\") pod \"redhat-marketplace-ltqgw\" (UID: \"44a897a0-3e5d-4c66-b159-a8830705e7cf\") " pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.510309 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44a897a0-3e5d-4c66-b159-a8830705e7cf-catalog-content\") pod \"redhat-marketplace-ltqgw\" (UID: \"44a897a0-3e5d-4c66-b159-a8830705e7cf\") " pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.510906 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44a897a0-3e5d-4c66-b159-a8830705e7cf-catalog-content\") pod \"redhat-marketplace-ltqgw\" (UID: \"44a897a0-3e5d-4c66-b159-a8830705e7cf\") " pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.511061 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44a897a0-3e5d-4c66-b159-a8830705e7cf-utilities\") pod \"redhat-marketplace-ltqgw\" (UID: \"44a897a0-3e5d-4c66-b159-a8830705e7cf\") " pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.537645 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgs7w\" (UniqueName: \"kubernetes.io/projected/44a897a0-3e5d-4c66-b159-a8830705e7cf-kube-api-access-hgs7w\") pod \"redhat-marketplace-ltqgw\" (UID: \"44a897a0-3e5d-4c66-b159-a8830705e7cf\") " pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.602121 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.612294 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4zbl6"] Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.628823 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.712637 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d87f2169-8a70-42de-aaf5-7728ff95fa50-config-volume\") pod \"d87f2169-8a70-42de-aaf5-7728ff95fa50\" (UID: \"d87f2169-8a70-42de-aaf5-7728ff95fa50\") " Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.713032 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggb4k\" (UniqueName: \"kubernetes.io/projected/d87f2169-8a70-42de-aaf5-7728ff95fa50-kube-api-access-ggb4k\") pod \"d87f2169-8a70-42de-aaf5-7728ff95fa50\" (UID: \"d87f2169-8a70-42de-aaf5-7728ff95fa50\") " Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.713082 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d87f2169-8a70-42de-aaf5-7728ff95fa50-secret-volume\") pod \"d87f2169-8a70-42de-aaf5-7728ff95fa50\" (UID: \"d87f2169-8a70-42de-aaf5-7728ff95fa50\") " Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.713883 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d87f2169-8a70-42de-aaf5-7728ff95fa50-config-volume" (OuterVolumeSpecName: "config-volume") pod "d87f2169-8a70-42de-aaf5-7728ff95fa50" (UID: "d87f2169-8a70-42de-aaf5-7728ff95fa50"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.715647 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.716177 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.719565 4814 patch_prober.go:28] interesting pod/console-f9d7485db-7x2qp container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.719640 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7x2qp" podUID="f487b7f4-c24b-431e-8868-a0285c8ed7bc" containerName="console" probeResult="failure" output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.720165 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d87f2169-8a70-42de-aaf5-7728ff95fa50-kube-api-access-ggb4k" (OuterVolumeSpecName: "kube-api-access-ggb4k") pod "d87f2169-8a70-42de-aaf5-7728ff95fa50" (UID: "d87f2169-8a70-42de-aaf5-7728ff95fa50"). InnerVolumeSpecName "kube-api-access-ggb4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.720418 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d87f2169-8a70-42de-aaf5-7728ff95fa50-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d87f2169-8a70-42de-aaf5-7728ff95fa50" (UID: "d87f2169-8a70-42de-aaf5-7728ff95fa50"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.759559 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:27:01 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:27:01 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:27:01 crc kubenswrapper[4814]: healthz check failed Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.759629 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.815166 4814 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d87f2169-8a70-42de-aaf5-7728ff95fa50-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.815207 4814 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d87f2169-8a70-42de-aaf5-7728ff95fa50-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.815222 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggb4k\" (UniqueName: \"kubernetes.io/projected/d87f2169-8a70-42de-aaf5-7728ff95fa50-kube-api-access-ggb4k\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.985334 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 27 16:27:01 crc kubenswrapper[4814]: E0227 16:27:01.985863 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d87f2169-8a70-42de-aaf5-7728ff95fa50" containerName="collect-profiles" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.985956 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="d87f2169-8a70-42de-aaf5-7728ff95fa50" containerName="collect-profiles" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.986137 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="d87f2169-8a70-42de-aaf5-7728ff95fa50" containerName="collect-profiles" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.986612 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.989044 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.989175 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 27 16:27:01 crc kubenswrapper[4814]: I0227 16:27:01.995859 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.006836 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.008311 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.018770 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.064703 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ltqgw"] Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.081040 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9ftgk"] Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.090008 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.093554 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9ftgk"] Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.099744 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.125814 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-5xv2z" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.127805 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f06ffc1-854e-46e5-8d4b-9c421af4e1dd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2f06ffc1-854e-46e5-8d4b-9c421af4e1dd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.128195 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f06ffc1-854e-46e5-8d4b-9c421af4e1dd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2f06ffc1-854e-46e5-8d4b-9c421af4e1dd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.127424 4814 patch_prober.go:28] interesting pod/downloads-7954f5f757-2n4tf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.131196 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2n4tf" podUID="98112972-5121-4d93-8490-5156a1a2f3ca" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.126501 4814 patch_prober.go:28] interesting pod/downloads-7954f5f757-2n4tf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.131624 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-2n4tf" podUID="98112972-5121-4d93-8490-5156a1a2f3ca" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.215423 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zbl6" event={"ID":"5df9c141-ab4f-40e3-b733-da2a2712a326","Type":"ContainerStarted","Data":"85ec660e523122f9e380a271b94600a98372d92e39c7c0ccca9cf657f8a08811"} Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.220931 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" event={"ID":"d87f2169-8a70-42de-aaf5-7728ff95fa50","Type":"ContainerDied","Data":"ae16799b7d5aaebd0d3a0366d13149ce400a26bbfd718fc2dd9891cab15194dd"} Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.221860 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae16799b7d5aaebd0d3a0366d13149ce400a26bbfd718fc2dd9891cab15194dd" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.222136 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.226141 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ltqgw" event={"ID":"44a897a0-3e5d-4c66-b159-a8830705e7cf","Type":"ContainerStarted","Data":"fd8f35f8dd59b93a9ad665b2251c172c40a2a828877279d623c095f7d218a603"} Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.235738 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b80cc2cd-d07d-4225-b078-9c70d8e189c7-catalog-content\") pod \"redhat-operators-9ftgk\" (UID: \"b80cc2cd-d07d-4225-b078-9c70d8e189c7\") " pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.238831 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f06ffc1-854e-46e5-8d4b-9c421af4e1dd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2f06ffc1-854e-46e5-8d4b-9c421af4e1dd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.239002 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f06ffc1-854e-46e5-8d4b-9c421af4e1dd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2f06ffc1-854e-46e5-8d4b-9c421af4e1dd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.239102 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b80cc2cd-d07d-4225-b078-9c70d8e189c7-utilities\") pod \"redhat-operators-9ftgk\" (UID: \"b80cc2cd-d07d-4225-b078-9c70d8e189c7\") " pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.239215 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kts7\" (UniqueName: \"kubernetes.io/projected/b80cc2cd-d07d-4225-b078-9c70d8e189c7-kube-api-access-9kts7\") pod \"redhat-operators-9ftgk\" (UID: \"b80cc2cd-d07d-4225-b078-9c70d8e189c7\") " pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.241039 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f06ffc1-854e-46e5-8d4b-9c421af4e1dd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2f06ffc1-854e-46e5-8d4b-9c421af4e1dd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.246485 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-pbsmc" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.261532 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f06ffc1-854e-46e5-8d4b-9c421af4e1dd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2f06ffc1-854e-46e5-8d4b-9c421af4e1dd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.309435 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.349369 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b80cc2cd-d07d-4225-b078-9c70d8e189c7-utilities\") pod \"redhat-operators-9ftgk\" (UID: \"b80cc2cd-d07d-4225-b078-9c70d8e189c7\") " pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.349480 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kts7\" (UniqueName: \"kubernetes.io/projected/b80cc2cd-d07d-4225-b078-9c70d8e189c7-kube-api-access-9kts7\") pod \"redhat-operators-9ftgk\" (UID: \"b80cc2cd-d07d-4225-b078-9c70d8e189c7\") " pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.352935 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b80cc2cd-d07d-4225-b078-9c70d8e189c7-catalog-content\") pod \"redhat-operators-9ftgk\" (UID: \"b80cc2cd-d07d-4225-b078-9c70d8e189c7\") " pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.353218 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b80cc2cd-d07d-4225-b078-9c70d8e189c7-utilities\") pod \"redhat-operators-9ftgk\" (UID: \"b80cc2cd-d07d-4225-b078-9c70d8e189c7\") " pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.353301 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b80cc2cd-d07d-4225-b078-9c70d8e189c7-catalog-content\") pod \"redhat-operators-9ftgk\" (UID: \"b80cc2cd-d07d-4225-b078-9c70d8e189c7\") " pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.395169 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kts7\" (UniqueName: \"kubernetes.io/projected/b80cc2cd-d07d-4225-b078-9c70d8e189c7-kube-api-access-9kts7\") pod \"redhat-operators-9ftgk\" (UID: \"b80cc2cd-d07d-4225-b078-9c70d8e189c7\") " pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.406105 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.490845 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dj4vx"] Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.492304 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.526735 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dj4vx"] Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.526854 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.584762 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tt9s" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.656830 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8781d31c-4701-4e6b-8065-495f1e890c9f-utilities\") pod \"redhat-operators-dj4vx\" (UID: \"8781d31c-4701-4e6b-8065-495f1e890c9f\") " pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.656873 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8781d31c-4701-4e6b-8065-495f1e890c9f-catalog-content\") pod \"redhat-operators-dj4vx\" (UID: \"8781d31c-4701-4e6b-8065-495f1e890c9f\") " pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.656909 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn5b7\" (UniqueName: \"kubernetes.io/projected/8781d31c-4701-4e6b-8065-495f1e890c9f-kube-api-access-zn5b7\") pod \"redhat-operators-dj4vx\" (UID: \"8781d31c-4701-4e6b-8065-495f1e890c9f\") " pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.688485 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 27 16:27:02 crc kubenswrapper[4814]: W0227 16:27:02.701904 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod2f06ffc1_854e_46e5_8d4b_9c421af4e1dd.slice/crio-ab8a29341e7ee9a43590bf4fa5423ab2daecbc9507548c99a0601e8ccb446be5 WatchSource:0}: Error finding container ab8a29341e7ee9a43590bf4fa5423ab2daecbc9507548c99a0601e8ccb446be5: Status 404 returned error can't find the container with id ab8a29341e7ee9a43590bf4fa5423ab2daecbc9507548c99a0601e8ccb446be5 Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.755745 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.757704 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn5b7\" (UniqueName: \"kubernetes.io/projected/8781d31c-4701-4e6b-8065-495f1e890c9f-kube-api-access-zn5b7\") pod \"redhat-operators-dj4vx\" (UID: \"8781d31c-4701-4e6b-8065-495f1e890c9f\") " pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.757868 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8781d31c-4701-4e6b-8065-495f1e890c9f-utilities\") pod \"redhat-operators-dj4vx\" (UID: \"8781d31c-4701-4e6b-8065-495f1e890c9f\") " pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.757894 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8781d31c-4701-4e6b-8065-495f1e890c9f-catalog-content\") pod \"redhat-operators-dj4vx\" (UID: \"8781d31c-4701-4e6b-8065-495f1e890c9f\") " pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.758374 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8781d31c-4701-4e6b-8065-495f1e890c9f-catalog-content\") pod \"redhat-operators-dj4vx\" (UID: \"8781d31c-4701-4e6b-8065-495f1e890c9f\") " pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.758999 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8781d31c-4701-4e6b-8065-495f1e890c9f-utilities\") pod \"redhat-operators-dj4vx\" (UID: \"8781d31c-4701-4e6b-8065-495f1e890c9f\") " pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.760051 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:27:02 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:27:02 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:27:02 crc kubenswrapper[4814]: healthz check failed Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.760086 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.779787 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn5b7\" (UniqueName: \"kubernetes.io/projected/8781d31c-4701-4e6b-8065-495f1e890c9f-kube-api-access-zn5b7\") pod \"redhat-operators-dj4vx\" (UID: \"8781d31c-4701-4e6b-8065-495f1e890c9f\") " pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.799588 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9ftgk"] Feb 27 16:27:02 crc kubenswrapper[4814]: W0227 16:27:02.844477 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb80cc2cd_d07d_4225_b078_9c70d8e189c7.slice/crio-498c9c32b480942bf3b30764e5cab7578b22e01bc4773bc209332f9e453d4b80 WatchSource:0}: Error finding container 498c9c32b480942bf3b30764e5cab7578b22e01bc4773bc209332f9e453d4b80: Status 404 returned error can't find the container with id 498c9c32b480942bf3b30764e5cab7578b22e01bc4773bc209332f9e453d4b80 Feb 27 16:27:02 crc kubenswrapper[4814]: I0227 16:27:02.849039 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.088425 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dj4vx"] Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.235793 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ftgk" event={"ID":"b80cc2cd-d07d-4225-b078-9c70d8e189c7","Type":"ContainerStarted","Data":"498c9c32b480942bf3b30764e5cab7578b22e01bc4773bc209332f9e453d4b80"} Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.239055 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2f06ffc1-854e-46e5-8d4b-9c421af4e1dd","Type":"ContainerStarted","Data":"ab8a29341e7ee9a43590bf4fa5423ab2daecbc9507548c99a0601e8ccb446be5"} Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.241232 4814 generic.go:334] "Generic (PLEG): container finished" podID="44a897a0-3e5d-4c66-b159-a8830705e7cf" containerID="11bac898d6d6afd051affa139d42a1071f6277d1c8310efe89949a7187f1a9ce" exitCode=0 Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.241355 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ltqgw" event={"ID":"44a897a0-3e5d-4c66-b159-a8830705e7cf","Type":"ContainerDied","Data":"11bac898d6d6afd051affa139d42a1071f6277d1c8310efe89949a7187f1a9ce"} Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.276769 4814 generic.go:334] "Generic (PLEG): container finished" podID="5df9c141-ab4f-40e3-b733-da2a2712a326" containerID="adf15ca9cd416ed78e0bd063bb50774a92c7dd0047a7cbd501b6805bf31e9f46" exitCode=0 Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.276868 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zbl6" event={"ID":"5df9c141-ab4f-40e3-b733-da2a2712a326","Type":"ContainerDied","Data":"adf15ca9cd416ed78e0bd063bb50774a92c7dd0047a7cbd501b6805bf31e9f46"} Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.353190 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.354044 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.357490 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.357752 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.379590 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.470615 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/533e678e-abf5-45d8-bb78-db2f3d7b7928-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"533e678e-abf5-45d8-bb78-db2f3d7b7928\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.470694 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/533e678e-abf5-45d8-bb78-db2f3d7b7928-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"533e678e-abf5-45d8-bb78-db2f3d7b7928\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.572373 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/533e678e-abf5-45d8-bb78-db2f3d7b7928-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"533e678e-abf5-45d8-bb78-db2f3d7b7928\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.572466 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/533e678e-abf5-45d8-bb78-db2f3d7b7928-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"533e678e-abf5-45d8-bb78-db2f3d7b7928\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.574955 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/533e678e-abf5-45d8-bb78-db2f3d7b7928-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"533e678e-abf5-45d8-bb78-db2f3d7b7928\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.607087 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/533e678e-abf5-45d8-bb78-db2f3d7b7928-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"533e678e-abf5-45d8-bb78-db2f3d7b7928\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.712878 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.760762 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:27:03 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:27:03 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:27:03 crc kubenswrapper[4814]: healthz check failed Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.760838 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:27:03 crc kubenswrapper[4814]: I0227 16:27:03.991482 4814 ???:1] "http: TLS handshake error from 192.168.126.11:33776: no serving certificate available for the kubelet" Feb 27 16:27:04 crc kubenswrapper[4814]: I0227 16:27:04.062621 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-sbf5t" Feb 27 16:27:04 crc kubenswrapper[4814]: I0227 16:27:04.295562 4814 generic.go:334] "Generic (PLEG): container finished" podID="2f06ffc1-854e-46e5-8d4b-9c421af4e1dd" containerID="3dd0b449c3565a502708854842358bdc4aef939fbda61129e9849c9ec06622d3" exitCode=0 Feb 27 16:27:04 crc kubenswrapper[4814]: I0227 16:27:04.295649 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2f06ffc1-854e-46e5-8d4b-9c421af4e1dd","Type":"ContainerDied","Data":"3dd0b449c3565a502708854842358bdc4aef939fbda61129e9849c9ec06622d3"} Feb 27 16:27:04 crc kubenswrapper[4814]: I0227 16:27:04.759292 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:27:04 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:27:04 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:27:04 crc kubenswrapper[4814]: healthz check failed Feb 27 16:27:04 crc kubenswrapper[4814]: I0227 16:27:04.759359 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:27:05 crc kubenswrapper[4814]: I0227 16:27:05.262119 4814 ???:1] "http: TLS handshake error from 192.168.126.11:33790: no serving certificate available for the kubelet" Feb 27 16:27:05 crc kubenswrapper[4814]: I0227 16:27:05.758744 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:27:05 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:27:05 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:27:05 crc kubenswrapper[4814]: healthz check failed Feb 27 16:27:05 crc kubenswrapper[4814]: I0227 16:27:05.758837 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:27:06 crc kubenswrapper[4814]: I0227 16:27:06.758280 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:27:06 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:27:06 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:27:06 crc kubenswrapper[4814]: healthz check failed Feb 27 16:27:06 crc kubenswrapper[4814]: I0227 16:27:06.758659 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:27:07 crc kubenswrapper[4814]: I0227 16:27:07.759600 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:27:07 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:27:07 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:27:07 crc kubenswrapper[4814]: healthz check failed Feb 27 16:27:07 crc kubenswrapper[4814]: I0227 16:27:07.759717 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:27:08 crc kubenswrapper[4814]: I0227 16:27:08.760496 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:27:08 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:27:08 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:27:08 crc kubenswrapper[4814]: healthz check failed Feb 27 16:27:08 crc kubenswrapper[4814]: I0227 16:27:08.761017 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:27:09 crc kubenswrapper[4814]: I0227 16:27:09.758440 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:27:09 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:27:09 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:27:09 crc kubenswrapper[4814]: healthz check failed Feb 27 16:27:09 crc kubenswrapper[4814]: I0227 16:27:09.758531 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:27:10 crc kubenswrapper[4814]: W0227 16:27:10.073604 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8781d31c_4701_4e6b_8065_495f1e890c9f.slice/crio-35ad2f22f8edc19859c225299b6e91f69a841b34b4031f20efa04b1ab40e607c WatchSource:0}: Error finding container 35ad2f22f8edc19859c225299b6e91f69a841b34b4031f20efa04b1ab40e607c: Status 404 returned error can't find the container with id 35ad2f22f8edc19859c225299b6e91f69a841b34b4031f20efa04b1ab40e607c Feb 27 16:27:10 crc kubenswrapper[4814]: I0227 16:27:10.212831 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 16:27:10 crc kubenswrapper[4814]: I0227 16:27:10.288947 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f06ffc1-854e-46e5-8d4b-9c421af4e1dd-kube-api-access\") pod \"2f06ffc1-854e-46e5-8d4b-9c421af4e1dd\" (UID: \"2f06ffc1-854e-46e5-8d4b-9c421af4e1dd\") " Feb 27 16:27:10 crc kubenswrapper[4814]: I0227 16:27:10.289033 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f06ffc1-854e-46e5-8d4b-9c421af4e1dd-kubelet-dir\") pod \"2f06ffc1-854e-46e5-8d4b-9c421af4e1dd\" (UID: \"2f06ffc1-854e-46e5-8d4b-9c421af4e1dd\") " Feb 27 16:27:10 crc kubenswrapper[4814]: I0227 16:27:10.289601 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f06ffc1-854e-46e5-8d4b-9c421af4e1dd-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2f06ffc1-854e-46e5-8d4b-9c421af4e1dd" (UID: "2f06ffc1-854e-46e5-8d4b-9c421af4e1dd"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:27:10 crc kubenswrapper[4814]: I0227 16:27:10.298378 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f06ffc1-854e-46e5-8d4b-9c421af4e1dd-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2f06ffc1-854e-46e5-8d4b-9c421af4e1dd" (UID: "2f06ffc1-854e-46e5-8d4b-9c421af4e1dd"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:27:10 crc kubenswrapper[4814]: I0227 16:27:10.343716 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dj4vx" event={"ID":"8781d31c-4701-4e6b-8065-495f1e890c9f","Type":"ContainerStarted","Data":"35ad2f22f8edc19859c225299b6e91f69a841b34b4031f20efa04b1ab40e607c"} Feb 27 16:27:10 crc kubenswrapper[4814]: I0227 16:27:10.345782 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2f06ffc1-854e-46e5-8d4b-9c421af4e1dd","Type":"ContainerDied","Data":"ab8a29341e7ee9a43590bf4fa5423ab2daecbc9507548c99a0601e8ccb446be5"} Feb 27 16:27:10 crc kubenswrapper[4814]: I0227 16:27:10.345816 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab8a29341e7ee9a43590bf4fa5423ab2daecbc9507548c99a0601e8ccb446be5" Feb 27 16:27:10 crc kubenswrapper[4814]: I0227 16:27:10.345881 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 16:27:10 crc kubenswrapper[4814]: I0227 16:27:10.391093 4814 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f06ffc1-854e-46e5-8d4b-9c421af4e1dd-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:10 crc kubenswrapper[4814]: I0227 16:27:10.391145 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f06ffc1-854e-46e5-8d4b-9c421af4e1dd-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:10 crc kubenswrapper[4814]: I0227 16:27:10.759813 4814 patch_prober.go:28] interesting pod/router-default-5444994796-wv7m2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 16:27:10 crc kubenswrapper[4814]: [-]has-synced failed: reason withheld Feb 27 16:27:10 crc kubenswrapper[4814]: [+]process-running ok Feb 27 16:27:10 crc kubenswrapper[4814]: healthz check failed Feb 27 16:27:10 crc kubenswrapper[4814]: I0227 16:27:10.760112 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wv7m2" podUID="13240f27-5cc2-4e90-9992-24f6beca4212" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 16:27:11 crc kubenswrapper[4814]: I0227 16:27:11.353568 4814 generic.go:334] "Generic (PLEG): container finished" podID="b80cc2cd-d07d-4225-b078-9c70d8e189c7" containerID="b5b291085889d8daaec79af7b1fcb2f89474bd83feeac9c571d9d17f26147820" exitCode=0 Feb 27 16:27:11 crc kubenswrapper[4814]: I0227 16:27:11.353665 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ftgk" event={"ID":"b80cc2cd-d07d-4225-b078-9c70d8e189c7","Type":"ContainerDied","Data":"b5b291085889d8daaec79af7b1fcb2f89474bd83feeac9c571d9d17f26147820"} Feb 27 16:27:11 crc kubenswrapper[4814]: I0227 16:27:11.717000 4814 patch_prober.go:28] interesting pod/console-f9d7485db-7x2qp container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Feb 27 16:27:11 crc kubenswrapper[4814]: I0227 16:27:11.717140 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7x2qp" podUID="f487b7f4-c24b-431e-8868-a0285c8ed7bc" containerName="console" probeResult="failure" output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" Feb 27 16:27:11 crc kubenswrapper[4814]: I0227 16:27:11.759374 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:27:11 crc kubenswrapper[4814]: I0227 16:27:11.763947 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-wv7m2" Feb 27 16:27:12 crc kubenswrapper[4814]: I0227 16:27:12.117666 4814 patch_prober.go:28] interesting pod/downloads-7954f5f757-2n4tf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Feb 27 16:27:12 crc kubenswrapper[4814]: I0227 16:27:12.117708 4814 patch_prober.go:28] interesting pod/downloads-7954f5f757-2n4tf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Feb 27 16:27:12 crc kubenswrapper[4814]: I0227 16:27:12.117768 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2n4tf" podUID="98112972-5121-4d93-8490-5156a1a2f3ca" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Feb 27 16:27:12 crc kubenswrapper[4814]: I0227 16:27:12.117780 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-2n4tf" podUID="98112972-5121-4d93-8490-5156a1a2f3ca" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Feb 27 16:27:14 crc kubenswrapper[4814]: I0227 16:27:14.047061 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs\") pod \"network-metrics-daemon-b7hx9\" (UID: \"45306fad-bea6-4c07-92ba-a910f8e26626\") " pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:27:14 crc kubenswrapper[4814]: I0227 16:27:14.050458 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 27 16:27:14 crc kubenswrapper[4814]: I0227 16:27:14.069135 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45306fad-bea6-4c07-92ba-a910f8e26626-metrics-certs\") pod \"network-metrics-daemon-b7hx9\" (UID: \"45306fad-bea6-4c07-92ba-a910f8e26626\") " pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:27:14 crc kubenswrapper[4814]: I0227 16:27:14.265323 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 27 16:27:14 crc kubenswrapper[4814]: I0227 16:27:14.270729 4814 ???:1] "http: TLS handshake error from 192.168.126.11:52996: no serving certificate available for the kubelet" Feb 27 16:27:14 crc kubenswrapper[4814]: I0227 16:27:14.273123 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b7hx9" Feb 27 16:27:15 crc kubenswrapper[4814]: I0227 16:27:15.814218 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-64967bf4f7-rmz25"] Feb 27 16:27:15 crc kubenswrapper[4814]: I0227 16:27:15.814476 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" podUID="7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9" containerName="controller-manager" containerID="cri-o://3ecfe545b52140b6f8d8e2da63d6dd1c7dbf2fe5f9fed179b29c2932932e24e2" gracePeriod=30 Feb 27 16:27:15 crc kubenswrapper[4814]: I0227 16:27:15.848452 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm"] Feb 27 16:27:15 crc kubenswrapper[4814]: I0227 16:27:15.848716 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" podUID="b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c" containerName="route-controller-manager" containerID="cri-o://526183fdafe1c67c9473802ba8555164094bdf32a27178105649c4bb6176725b" gracePeriod=30 Feb 27 16:27:16 crc kubenswrapper[4814]: I0227 16:27:16.386776 4814 generic.go:334] "Generic (PLEG): container finished" podID="b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c" containerID="526183fdafe1c67c9473802ba8555164094bdf32a27178105649c4bb6176725b" exitCode=0 Feb 27 16:27:16 crc kubenswrapper[4814]: I0227 16:27:16.386832 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" event={"ID":"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c","Type":"ContainerDied","Data":"526183fdafe1c67c9473802ba8555164094bdf32a27178105649c4bb6176725b"} Feb 27 16:27:17 crc kubenswrapper[4814]: I0227 16:27:17.396552 4814 generic.go:334] "Generic (PLEG): container finished" podID="7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9" containerID="3ecfe545b52140b6f8d8e2da63d6dd1c7dbf2fe5f9fed179b29c2932932e24e2" exitCode=0 Feb 27 16:27:17 crc kubenswrapper[4814]: I0227 16:27:17.396633 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" event={"ID":"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9","Type":"ContainerDied","Data":"3ecfe545b52140b6f8d8e2da63d6dd1c7dbf2fe5f9fed179b29c2932932e24e2"} Feb 27 16:27:19 crc kubenswrapper[4814]: I0227 16:27:19.803887 4814 patch_prober.go:28] interesting pod/route-controller-manager-687b4d5955-4wznm container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.46:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 16:27:19 crc kubenswrapper[4814]: I0227 16:27:19.804200 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" podUID="b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.46:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 16:27:20 crc kubenswrapper[4814]: I0227 16:27:20.116939 4814 patch_prober.go:28] interesting pod/controller-manager-64967bf4f7-rmz25 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.45:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 16:27:20 crc kubenswrapper[4814]: I0227 16:27:20.117002 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" podUID="7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.45:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 16:27:20 crc kubenswrapper[4814]: I0227 16:27:20.240902 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:27:21 crc kubenswrapper[4814]: I0227 16:27:21.723619 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:27:21 crc kubenswrapper[4814]: I0227 16:27:21.736336 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:27:22 crc kubenswrapper[4814]: I0227 16:27:22.139059 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-2n4tf" Feb 27 16:27:22 crc kubenswrapper[4814]: I0227 16:27:22.902376 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:27:22 crc kubenswrapper[4814]: I0227 16:27:22.902485 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.444749 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" event={"ID":"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c","Type":"ContainerDied","Data":"fc1fec091a5ae07f6b635962e96f928452d581369755e5690ea05ff1a08e72e0"} Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.445379 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc1fec091a5ae07f6b635962e96f928452d581369755e5690ea05ff1a08e72e0" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.447291 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" event={"ID":"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9","Type":"ContainerDied","Data":"213dc6cee9532429c439ee6d28414db5c00acf5519c9ef6c258e764bf8efe044"} Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.447323 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="213dc6cee9532429c439ee6d28414db5c00acf5519c9ef6c258e764bf8efe044" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.484248 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.489192 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.514445 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-serving-cert\") pod \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.514535 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-config\") pod \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.514592 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-client-ca\") pod \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.514674 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-proxy-ca-bundles\") pod \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.514717 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r68sl\" (UniqueName: \"kubernetes.io/projected/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-kube-api-access-r68sl\") pod \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\" (UID: \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\") " Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.516226 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9" (UID: "7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.516807 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-client-ca" (OuterVolumeSpecName: "client-ca") pod "7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9" (UID: "7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.516832 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-config" (OuterVolumeSpecName: "config") pod "7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9" (UID: "7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.525126 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-kube-api-access-r68sl" (OuterVolumeSpecName: "kube-api-access-r68sl") pod "b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c" (UID: "b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c"). InnerVolumeSpecName "kube-api-access-r68sl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.534876 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6"] Feb 27 16:27:25 crc kubenswrapper[4814]: E0227 16:27:25.535550 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f06ffc1-854e-46e5-8d4b-9c421af4e1dd" containerName="pruner" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.535613 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f06ffc1-854e-46e5-8d4b-9c421af4e1dd" containerName="pruner" Feb 27 16:27:25 crc kubenswrapper[4814]: E0227 16:27:25.535637 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c" containerName="route-controller-manager" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.535654 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c" containerName="route-controller-manager" Feb 27 16:27:25 crc kubenswrapper[4814]: E0227 16:27:25.535715 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9" containerName="controller-manager" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.535729 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9" containerName="controller-manager" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.535912 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c" containerName="route-controller-manager" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.535944 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f06ffc1-854e-46e5-8d4b-9c421af4e1dd" containerName="pruner" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.535963 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9" containerName="controller-manager" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.536821 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.543152 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6"] Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.544922 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9" (UID: "7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.616086 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-client-ca\") pod \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\" (UID: \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\") " Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.616142 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrqsj\" (UniqueName: \"kubernetes.io/projected/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-kube-api-access-hrqsj\") pod \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\" (UID: \"7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9\") " Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.616826 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-config\") pod \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\" (UID: \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\") " Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.616892 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-serving-cert\") pod \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\" (UID: \"b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c\") " Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.617049 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-serving-cert\") pod \"route-controller-manager-869499dcbc-sq9h6\" (UID: \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\") " pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.617120 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chlpj\" (UniqueName: \"kubernetes.io/projected/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-kube-api-access-chlpj\") pod \"route-controller-manager-869499dcbc-sq9h6\" (UID: \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\") " pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.617169 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-client-ca\") pod \"route-controller-manager-869499dcbc-sq9h6\" (UID: \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\") " pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.617214 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-config\") pod \"route-controller-manager-869499dcbc-sq9h6\" (UID: \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\") " pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.617412 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r68sl\" (UniqueName: \"kubernetes.io/projected/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-kube-api-access-r68sl\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.617434 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.617453 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.617468 4814 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.617483 4814 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.617566 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-config" (OuterVolumeSpecName: "config") pod "b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c" (UID: "b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.618791 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-client-ca" (OuterVolumeSpecName: "client-ca") pod "b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c" (UID: "b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.619855 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-kube-api-access-hrqsj" (OuterVolumeSpecName: "kube-api-access-hrqsj") pod "7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9" (UID: "7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9"). InnerVolumeSpecName "kube-api-access-hrqsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.623871 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c" (UID: "b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.717872 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chlpj\" (UniqueName: \"kubernetes.io/projected/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-kube-api-access-chlpj\") pod \"route-controller-manager-869499dcbc-sq9h6\" (UID: \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\") " pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.717928 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-client-ca\") pod \"route-controller-manager-869499dcbc-sq9h6\" (UID: \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\") " pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.717967 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-config\") pod \"route-controller-manager-869499dcbc-sq9h6\" (UID: \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\") " pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.718041 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-serving-cert\") pod \"route-controller-manager-869499dcbc-sq9h6\" (UID: \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\") " pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.718097 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.718107 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.718117 4814 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.718128 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrqsj\" (UniqueName: \"kubernetes.io/projected/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9-kube-api-access-hrqsj\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.719396 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-client-ca\") pod \"route-controller-manager-869499dcbc-sq9h6\" (UID: \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\") " pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.720345 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-config\") pod \"route-controller-manager-869499dcbc-sq9h6\" (UID: \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\") " pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.723731 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-serving-cert\") pod \"route-controller-manager-869499dcbc-sq9h6\" (UID: \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\") " pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.735954 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chlpj\" (UniqueName: \"kubernetes.io/projected/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-kube-api-access-chlpj\") pod \"route-controller-manager-869499dcbc-sq9h6\" (UID: \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\") " pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:25 crc kubenswrapper[4814]: I0227 16:27:25.884809 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:26 crc kubenswrapper[4814]: I0227 16:27:26.452460 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64967bf4f7-rmz25" Feb 27 16:27:26 crc kubenswrapper[4814]: I0227 16:27:26.452481 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm" Feb 27 16:27:26 crc kubenswrapper[4814]: I0227 16:27:26.502090 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-64967bf4f7-rmz25"] Feb 27 16:27:26 crc kubenswrapper[4814]: I0227 16:27:26.502141 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-64967bf4f7-rmz25"] Feb 27 16:27:26 crc kubenswrapper[4814]: I0227 16:27:26.513938 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm"] Feb 27 16:27:26 crc kubenswrapper[4814]: I0227 16:27:26.519372 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-687b4d5955-4wznm"] Feb 27 16:27:27 crc kubenswrapper[4814]: E0227 16:27:27.125492 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 16:27:27 crc kubenswrapper[4814]: E0227 16:27:27.125731 4814 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 16:27:27 crc kubenswrapper[4814]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 16:27:27 crc kubenswrapper[4814]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8w6zl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536826-dgwn4_openshift-infra(95913cb4-39f1-44c0-ac49-0a2d51047679): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Feb 27 16:27:27 crc kubenswrapper[4814]: > logger="UnhandledError" Feb 27 16:27:27 crc kubenswrapper[4814]: E0227 16:27:27.126954 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29536826-dgwn4" podUID="95913cb4-39f1-44c0-ac49-0a2d51047679" Feb 27 16:27:27 crc kubenswrapper[4814]: E0227 16:27:27.459960 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536826-dgwn4" podUID="95913cb4-39f1-44c0-ac49-0a2d51047679" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.495212 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9" path="/var/lib/kubelet/pods/7bb4bfd0-e2cf-4dd9-b0b8-75ca651d98a9/volumes" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.496698 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c" path="/var/lib/kubelet/pods/b271c61a-ca89-4e44-91ce-5f5b4c6a7d5c/volumes" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.501807 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd"] Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.502775 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.505340 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.505547 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.506064 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.506219 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.506764 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.506948 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.518882 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.523228 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd"] Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.663241 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-proxy-ca-bundles\") pod \"controller-manager-6f9f8b7999-tmwhd\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.663555 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-config\") pod \"controller-manager-6f9f8b7999-tmwhd\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.663633 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-client-ca\") pod \"controller-manager-6f9f8b7999-tmwhd\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.663662 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrj5c\" (UniqueName: \"kubernetes.io/projected/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-kube-api-access-lrj5c\") pod \"controller-manager-6f9f8b7999-tmwhd\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.663692 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-serving-cert\") pod \"controller-manager-6f9f8b7999-tmwhd\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.764821 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-proxy-ca-bundles\") pod \"controller-manager-6f9f8b7999-tmwhd\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.765001 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-config\") pod \"controller-manager-6f9f8b7999-tmwhd\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.765071 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-client-ca\") pod \"controller-manager-6f9f8b7999-tmwhd\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.765106 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrj5c\" (UniqueName: \"kubernetes.io/projected/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-kube-api-access-lrj5c\") pod \"controller-manager-6f9f8b7999-tmwhd\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.765148 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-serving-cert\") pod \"controller-manager-6f9f8b7999-tmwhd\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.766286 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-proxy-ca-bundles\") pod \"controller-manager-6f9f8b7999-tmwhd\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.766487 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-client-ca\") pod \"controller-manager-6f9f8b7999-tmwhd\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.767579 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-config\") pod \"controller-manager-6f9f8b7999-tmwhd\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.773346 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-serving-cert\") pod \"controller-manager-6f9f8b7999-tmwhd\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.782139 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrj5c\" (UniqueName: \"kubernetes.io/projected/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-kube-api-access-lrj5c\") pod \"controller-manager-6f9f8b7999-tmwhd\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:28 crc kubenswrapper[4814]: I0227 16:27:28.835022 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:31 crc kubenswrapper[4814]: I0227 16:27:31.119728 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 27 16:27:32 crc kubenswrapper[4814]: I0227 16:27:32.235304 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6c5hj" Feb 27 16:27:33 crc kubenswrapper[4814]: I0227 16:27:33.336733 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 27 16:27:33 crc kubenswrapper[4814]: I0227 16:27:33.337930 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 16:27:33 crc kubenswrapper[4814]: I0227 16:27:33.350520 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 27 16:27:33 crc kubenswrapper[4814]: I0227 16:27:33.450014 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9543a634-2567-4f4e-8e45-5c2eccb2344e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9543a634-2567-4f4e-8e45-5c2eccb2344e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 16:27:33 crc kubenswrapper[4814]: I0227 16:27:33.450080 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9543a634-2567-4f4e-8e45-5c2eccb2344e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9543a634-2567-4f4e-8e45-5c2eccb2344e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 16:27:33 crc kubenswrapper[4814]: I0227 16:27:33.551581 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9543a634-2567-4f4e-8e45-5c2eccb2344e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9543a634-2567-4f4e-8e45-5c2eccb2344e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 16:27:33 crc kubenswrapper[4814]: I0227 16:27:33.551649 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9543a634-2567-4f4e-8e45-5c2eccb2344e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9543a634-2567-4f4e-8e45-5c2eccb2344e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 16:27:33 crc kubenswrapper[4814]: I0227 16:27:33.552195 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9543a634-2567-4f4e-8e45-5c2eccb2344e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9543a634-2567-4f4e-8e45-5c2eccb2344e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 16:27:33 crc kubenswrapper[4814]: I0227 16:27:33.585307 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9543a634-2567-4f4e-8e45-5c2eccb2344e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9543a634-2567-4f4e-8e45-5c2eccb2344e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 16:27:33 crc kubenswrapper[4814]: I0227 16:27:33.673557 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 16:27:35 crc kubenswrapper[4814]: I0227 16:27:35.831571 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd"] Feb 27 16:27:35 crc kubenswrapper[4814]: I0227 16:27:35.911020 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6"] Feb 27 16:27:37 crc kubenswrapper[4814]: I0227 16:27:37.360823 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 27 16:27:37 crc kubenswrapper[4814]: I0227 16:27:37.362512 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 27 16:27:37 crc kubenswrapper[4814]: I0227 16:27:37.367138 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 27 16:27:37 crc kubenswrapper[4814]: I0227 16:27:37.436765 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/306421d3-af6e-43c8-9411-a8eaaf583862-kubelet-dir\") pod \"installer-9-crc\" (UID: \"306421d3-af6e-43c8-9411-a8eaaf583862\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 16:27:37 crc kubenswrapper[4814]: I0227 16:27:37.436955 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/306421d3-af6e-43c8-9411-a8eaaf583862-var-lock\") pod \"installer-9-crc\" (UID: \"306421d3-af6e-43c8-9411-a8eaaf583862\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 16:27:37 crc kubenswrapper[4814]: I0227 16:27:37.437177 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/306421d3-af6e-43c8-9411-a8eaaf583862-kube-api-access\") pod \"installer-9-crc\" (UID: \"306421d3-af6e-43c8-9411-a8eaaf583862\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 16:27:37 crc kubenswrapper[4814]: I0227 16:27:37.539368 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/306421d3-af6e-43c8-9411-a8eaaf583862-kubelet-dir\") pod \"installer-9-crc\" (UID: \"306421d3-af6e-43c8-9411-a8eaaf583862\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 16:27:37 crc kubenswrapper[4814]: I0227 16:27:37.539551 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/306421d3-af6e-43c8-9411-a8eaaf583862-var-lock\") pod \"installer-9-crc\" (UID: \"306421d3-af6e-43c8-9411-a8eaaf583862\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 16:27:37 crc kubenswrapper[4814]: I0227 16:27:37.539591 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/306421d3-af6e-43c8-9411-a8eaaf583862-kubelet-dir\") pod \"installer-9-crc\" (UID: \"306421d3-af6e-43c8-9411-a8eaaf583862\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 16:27:37 crc kubenswrapper[4814]: I0227 16:27:37.539635 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/306421d3-af6e-43c8-9411-a8eaaf583862-kube-api-access\") pod \"installer-9-crc\" (UID: \"306421d3-af6e-43c8-9411-a8eaaf583862\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 16:27:37 crc kubenswrapper[4814]: I0227 16:27:37.539704 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/306421d3-af6e-43c8-9411-a8eaaf583862-var-lock\") pod \"installer-9-crc\" (UID: \"306421d3-af6e-43c8-9411-a8eaaf583862\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 16:27:37 crc kubenswrapper[4814]: I0227 16:27:37.581545 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/306421d3-af6e-43c8-9411-a8eaaf583862-kube-api-access\") pod \"installer-9-crc\" (UID: \"306421d3-af6e-43c8-9411-a8eaaf583862\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 16:27:37 crc kubenswrapper[4814]: I0227 16:27:37.693307 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 27 16:27:38 crc kubenswrapper[4814]: E0227 16:27:38.623408 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 16:27:38 crc kubenswrapper[4814]: E0227 16:27:38.623652 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tpb22,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-4zbl6_openshift-marketplace(5df9c141-ab4f-40e3-b733-da2a2712a326): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 16:27:38 crc kubenswrapper[4814]: E0227 16:27:38.624889 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-4zbl6" podUID="5df9c141-ab4f-40e3-b733-da2a2712a326" Feb 27 16:27:38 crc kubenswrapper[4814]: E0227 16:27:38.658426 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 16:27:38 crc kubenswrapper[4814]: E0227 16:27:38.658667 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h5cld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-cm8p5_openshift-marketplace(f1ed29a3-a2f5-46e1-ac86-b3adc8c52375): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 16:27:38 crc kubenswrapper[4814]: E0227 16:27:38.659954 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-cm8p5" podUID="f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" Feb 27 16:27:40 crc kubenswrapper[4814]: E0227 16:27:40.073418 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-4zbl6" podUID="5df9c141-ab4f-40e3-b733-da2a2712a326" Feb 27 16:27:40 crc kubenswrapper[4814]: E0227 16:27:40.074534 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-cm8p5" podUID="f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" Feb 27 16:27:40 crc kubenswrapper[4814]: E0227 16:27:40.130322 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 16:27:40 crc kubenswrapper[4814]: E0227 16:27:40.130544 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vgd64,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-sk6rs_openshift-marketplace(41b851a0-9bda-4e20-a692-5f796f55a749): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 16:27:40 crc kubenswrapper[4814]: E0227 16:27:40.131745 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-sk6rs" podUID="41b851a0-9bda-4e20-a692-5f796f55a749" Feb 27 16:27:40 crc kubenswrapper[4814]: E0227 16:27:40.201401 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 16:27:40 crc kubenswrapper[4814]: E0227 16:27:40.202058 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rddgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-c5wns_openshift-marketplace(d3a647cf-eb96-40b8-982c-8e8b869e57c4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 16:27:40 crc kubenswrapper[4814]: E0227 16:27:40.204165 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-c5wns" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" Feb 27 16:27:40 crc kubenswrapper[4814]: I0227 16:27:40.516741 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-b7hx9"] Feb 27 16:27:40 crc kubenswrapper[4814]: I0227 16:27:40.561470 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"533e678e-abf5-45d8-bb78-db2f3d7b7928","Type":"ContainerStarted","Data":"93b3b5312df6b856f944d69621378a67ba4337ae0faa8d52cd7f6fb4faa77ab9"} Feb 27 16:27:44 crc kubenswrapper[4814]: E0227 16:27:44.068694 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-sk6rs" podUID="41b851a0-9bda-4e20-a692-5f796f55a749" Feb 27 16:27:44 crc kubenswrapper[4814]: E0227 16:27:44.069305 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-c5wns" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" Feb 27 16:27:44 crc kubenswrapper[4814]: E0227 16:27:44.079532 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 16:27:44 crc kubenswrapper[4814]: E0227 16:27:44.079858 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9kts7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-9ftgk_openshift-marketplace(b80cc2cd-d07d-4225-b078-9c70d8e189c7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 16:27:44 crc kubenswrapper[4814]: E0227 16:27:44.081574 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-9ftgk" podUID="b80cc2cd-d07d-4225-b078-9c70d8e189c7" Feb 27 16:27:44 crc kubenswrapper[4814]: W0227 16:27:44.122590 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45306fad_bea6_4c07_92ba_a910f8e26626.slice/crio-d7ff00f6bdf67cbcbb2a240115da3bbc4450be190633c76580680c590cd0655d WatchSource:0}: Error finding container d7ff00f6bdf67cbcbb2a240115da3bbc4450be190633c76580680c590cd0655d: Status 404 returned error can't find the container with id d7ff00f6bdf67cbcbb2a240115da3bbc4450be190633c76580680c590cd0655d Feb 27 16:27:44 crc kubenswrapper[4814]: I0227 16:27:44.547002 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd"] Feb 27 16:27:44 crc kubenswrapper[4814]: W0227 16:27:44.564288 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25dfc3d0_6a18_48fd_9fcf_68313673e6b7.slice/crio-59030b77db0b836e296f9e308578691b14c4c6dd9feec56eb747b4d25a56cd6b WatchSource:0}: Error finding container 59030b77db0b836e296f9e308578691b14c4c6dd9feec56eb747b4d25a56cd6b: Status 404 returned error can't find the container with id 59030b77db0b836e296f9e308578691b14c4c6dd9feec56eb747b4d25a56cd6b Feb 27 16:27:44 crc kubenswrapper[4814]: I0227 16:27:44.592074 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ltqgw" event={"ID":"44a897a0-3e5d-4c66-b159-a8830705e7cf","Type":"ContainerStarted","Data":"bfe8a1cc35dcf3fd440b085311f857ee97d33222ba4804011c9850b7554e4f3e"} Feb 27 16:27:44 crc kubenswrapper[4814]: I0227 16:27:44.600837 4814 generic.go:334] "Generic (PLEG): container finished" podID="8781d31c-4701-4e6b-8065-495f1e890c9f" containerID="dc5ad3d7ab001b79bec729342ae0f4e69e413119cb592d087f67b00d0eae5d68" exitCode=0 Feb 27 16:27:44 crc kubenswrapper[4814]: I0227 16:27:44.600947 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dj4vx" event={"ID":"8781d31c-4701-4e6b-8065-495f1e890c9f","Type":"ContainerDied","Data":"dc5ad3d7ab001b79bec729342ae0f4e69e413119cb592d087f67b00d0eae5d68"} Feb 27 16:27:44 crc kubenswrapper[4814]: I0227 16:27:44.602584 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6"] Feb 27 16:27:44 crc kubenswrapper[4814]: I0227 16:27:44.628294 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" event={"ID":"25dfc3d0-6a18-48fd-9fcf-68313673e6b7","Type":"ContainerStarted","Data":"59030b77db0b836e296f9e308578691b14c4c6dd9feec56eb747b4d25a56cd6b"} Feb 27 16:27:44 crc kubenswrapper[4814]: I0227 16:27:44.659679 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 27 16:27:44 crc kubenswrapper[4814]: I0227 16:27:44.665888 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" event={"ID":"45306fad-bea6-4c07-92ba-a910f8e26626","Type":"ContainerStarted","Data":"6f6daeefc7a730a6ee8ad58ef0cddfab4d8deaa838faa7825b6b0b5e3ddb269a"} Feb 27 16:27:44 crc kubenswrapper[4814]: I0227 16:27:44.665940 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" event={"ID":"45306fad-bea6-4c07-92ba-a910f8e26626","Type":"ContainerStarted","Data":"d7ff00f6bdf67cbcbb2a240115da3bbc4450be190633c76580680c590cd0655d"} Feb 27 16:27:44 crc kubenswrapper[4814]: W0227 16:27:44.673986 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod9543a634_2567_4f4e_8e45_5c2eccb2344e.slice/crio-161b234a09c9e0ed136413eddceaa8df01cc3511000699f6189891733ef0c00b WatchSource:0}: Error finding container 161b234a09c9e0ed136413eddceaa8df01cc3511000699f6189891733ef0c00b: Status 404 returned error can't find the container with id 161b234a09c9e0ed136413eddceaa8df01cc3511000699f6189891733ef0c00b Feb 27 16:27:44 crc kubenswrapper[4814]: I0227 16:27:44.679696 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-972lm" event={"ID":"2cdfdb15-014d-40fc-9a7b-a905278f1249","Type":"ContainerStarted","Data":"eecc9514e37f521f316d91ee6da545090202782ae23267acbcd118ea880c0edd"} Feb 27 16:27:44 crc kubenswrapper[4814]: E0227 16:27:44.712416 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-9ftgk" podUID="b80cc2cd-d07d-4225-b078-9c70d8e189c7" Feb 27 16:27:44 crc kubenswrapper[4814]: I0227 16:27:44.764383 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.591904 4814 csr.go:261] certificate signing request csr-9rb9b is approved, waiting to be issued Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.603965 4814 csr.go:257] certificate signing request csr-9rb9b is issued Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.687920 4814 generic.go:334] "Generic (PLEG): container finished" podID="44a897a0-3e5d-4c66-b159-a8830705e7cf" containerID="bfe8a1cc35dcf3fd440b085311f857ee97d33222ba4804011c9850b7554e4f3e" exitCode=0 Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.687962 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ltqgw" event={"ID":"44a897a0-3e5d-4c66-b159-a8830705e7cf","Type":"ContainerDied","Data":"bfe8a1cc35dcf3fd440b085311f857ee97d33222ba4804011c9850b7554e4f3e"} Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.691646 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dj4vx" event={"ID":"8781d31c-4701-4e6b-8065-495f1e890c9f","Type":"ContainerStarted","Data":"f0d8a4412d8e779ac5926ee3f2fdf5680097423b4072d7ce9b4e0d49d5a87a8f"} Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.693431 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" podUID="25dfc3d0-6a18-48fd-9fcf-68313673e6b7" containerName="controller-manager" containerID="cri-o://2147b1f6d7df54d96f3b06111bef8a1dec30c4592cc9ec706bef48ab0ef1852e" gracePeriod=30 Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.693517 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" event={"ID":"25dfc3d0-6a18-48fd-9fcf-68313673e6b7","Type":"ContainerStarted","Data":"2147b1f6d7df54d96f3b06111bef8a1dec30c4592cc9ec706bef48ab0ef1852e"} Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.693842 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.699069 4814 generic.go:334] "Generic (PLEG): container finished" podID="2cdfdb15-014d-40fc-9a7b-a905278f1249" containerID="eecc9514e37f521f316d91ee6da545090202782ae23267acbcd118ea880c0edd" exitCode=0 Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.699151 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-972lm" event={"ID":"2cdfdb15-014d-40fc-9a7b-a905278f1249","Type":"ContainerDied","Data":"eecc9514e37f521f316d91ee6da545090202782ae23267acbcd118ea880c0edd"} Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.702403 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.704033 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b7hx9" event={"ID":"45306fad-bea6-4c07-92ba-a910f8e26626","Type":"ContainerStarted","Data":"a36e50767e0ca2beff73ee22e56fa5a08b4fb75dbcd4b2fa560ecbe53a033367"} Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.711880 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"306421d3-af6e-43c8-9411-a8eaaf583862","Type":"ContainerStarted","Data":"84352145e637126e0a3480c234b3eae3bd5f7bc9152740430bf73092114a1a90"} Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.711960 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"306421d3-af6e-43c8-9411-a8eaaf583862","Type":"ContainerStarted","Data":"159b163211cc0919ef75fa57e0135ef1a7e25179bbb9a1b69ef98058200a8e7e"} Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.715057 4814 generic.go:334] "Generic (PLEG): container finished" podID="95913cb4-39f1-44c0-ac49-0a2d51047679" containerID="9a038ee1f7a4efc09ec1a2ebfb4250554cbbcea6f81b89bc6e1317e43f091d3c" exitCode=0 Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.715141 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536826-dgwn4" event={"ID":"95913cb4-39f1-44c0-ac49-0a2d51047679","Type":"ContainerDied","Data":"9a038ee1f7a4efc09ec1a2ebfb4250554cbbcea6f81b89bc6e1317e43f091d3c"} Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.717243 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" event={"ID":"a6b20452-fd7b-4169-8e38-d5405fa0cbb3","Type":"ContainerStarted","Data":"e9eb5c9010c110dcd4eacb25fd7c74e465f10e847e8fd9ca9e8c4c8c0fb3b402"} Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.717312 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" event={"ID":"a6b20452-fd7b-4169-8e38-d5405fa0cbb3","Type":"ContainerStarted","Data":"5939f69cfd7643e39ea507709496007b5cfef7f9e1890a890f222fb71cd6222e"} Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.717391 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" podUID="a6b20452-fd7b-4169-8e38-d5405fa0cbb3" containerName="route-controller-manager" containerID="cri-o://e9eb5c9010c110dcd4eacb25fd7c74e465f10e847e8fd9ca9e8c4c8c0fb3b402" gracePeriod=30 Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.717444 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.721311 4814 generic.go:334] "Generic (PLEG): container finished" podID="533e678e-abf5-45d8-bb78-db2f3d7b7928" containerID="e9c7e201fedbf949bcc1f50cd5a3ed0c1b74403c621dd8949e2f8ef037fda116" exitCode=0 Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.721379 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"533e678e-abf5-45d8-bb78-db2f3d7b7928","Type":"ContainerDied","Data":"e9c7e201fedbf949bcc1f50cd5a3ed0c1b74403c621dd8949e2f8ef037fda116"} Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.724345 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.725166 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9543a634-2567-4f4e-8e45-5c2eccb2344e","Type":"ContainerStarted","Data":"a9e458d9c2722ee4d535cc7b7040fff5ed2931f762d4313e4878afe2c9335015"} Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.725192 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9543a634-2567-4f4e-8e45-5c2eccb2344e","Type":"ContainerStarted","Data":"161b234a09c9e0ed136413eddceaa8df01cc3511000699f6189891733ef0c00b"} Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.768349 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-b7hx9" podStartSLOduration=212.768327241 podStartE2EDuration="3m32.768327241s" podCreationTimestamp="2026-02-27 16:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:27:45.766454797 +0000 UTC m=+278.219079627" watchObservedRunningTime="2026-02-27 16:27:45.768327241 +0000 UTC m=+278.220952071" Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.791375 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" podStartSLOduration=30.791351027 podStartE2EDuration="30.791351027s" podCreationTimestamp="2026-02-27 16:27:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:27:45.791273164 +0000 UTC m=+278.243897984" watchObservedRunningTime="2026-02-27 16:27:45.791351027 +0000 UTC m=+278.243975857" Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.814529 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=12.814505778000001 podStartE2EDuration="12.814505778s" podCreationTimestamp="2026-02-27 16:27:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:27:45.812709157 +0000 UTC m=+278.265333997" watchObservedRunningTime="2026-02-27 16:27:45.814505778 +0000 UTC m=+278.267130608" Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.837415 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" podStartSLOduration=30.83739713 podStartE2EDuration="30.83739713s" podCreationTimestamp="2026-02-27 16:27:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:27:45.831193517 +0000 UTC m=+278.283818337" watchObservedRunningTime="2026-02-27 16:27:45.83739713 +0000 UTC m=+278.290021960" Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.885045 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=8.885021456 podStartE2EDuration="8.885021456s" podCreationTimestamp="2026-02-27 16:27:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:27:45.883892917 +0000 UTC m=+278.336517747" watchObservedRunningTime="2026-02-27 16:27:45.885021456 +0000 UTC m=+278.337646276" Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.888331 4814 patch_prober.go:28] interesting pod/route-controller-manager-869499dcbc-sq9h6 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" start-of-body= Feb 27 16:27:45 crc kubenswrapper[4814]: I0227 16:27:45.888398 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" podUID="a6b20452-fd7b-4169-8e38-d5405fa0cbb3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.128327 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.132749 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.158184 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-565994456b-sh4t6"] Feb 27 16:27:46 crc kubenswrapper[4814]: E0227 16:27:46.158492 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6b20452-fd7b-4169-8e38-d5405fa0cbb3" containerName="route-controller-manager" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.158506 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6b20452-fd7b-4169-8e38-d5405fa0cbb3" containerName="route-controller-manager" Feb 27 16:27:46 crc kubenswrapper[4814]: E0227 16:27:46.158525 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25dfc3d0-6a18-48fd-9fcf-68313673e6b7" containerName="controller-manager" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.158531 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="25dfc3d0-6a18-48fd-9fcf-68313673e6b7" containerName="controller-manager" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.158634 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="25dfc3d0-6a18-48fd-9fcf-68313673e6b7" containerName="controller-manager" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.158650 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6b20452-fd7b-4169-8e38-d5405fa0cbb3" containerName="route-controller-manager" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.159069 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.181555 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-565994456b-sh4t6"] Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.275327 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-proxy-ca-bundles\") pod \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.275429 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrj5c\" (UniqueName: \"kubernetes.io/projected/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-kube-api-access-lrj5c\") pod \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.275460 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-config\") pod \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\" (UID: \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\") " Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.275507 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-client-ca\") pod \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\" (UID: \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\") " Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.275577 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-config\") pod \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.275609 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chlpj\" (UniqueName: \"kubernetes.io/projected/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-kube-api-access-chlpj\") pod \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\" (UID: \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\") " Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.275629 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-serving-cert\") pod \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.275679 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-client-ca\") pod \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\" (UID: \"25dfc3d0-6a18-48fd-9fcf-68313673e6b7\") " Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.275714 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-serving-cert\") pod \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\" (UID: \"a6b20452-fd7b-4169-8e38-d5405fa0cbb3\") " Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.275928 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-config\") pod \"controller-manager-565994456b-sh4t6\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.275975 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-serving-cert\") pod \"controller-manager-565994456b-sh4t6\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.276002 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kqn8\" (UniqueName: \"kubernetes.io/projected/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-kube-api-access-7kqn8\") pod \"controller-manager-565994456b-sh4t6\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.276086 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-client-ca\") pod \"controller-manager-565994456b-sh4t6\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.276104 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-proxy-ca-bundles\") pod \"controller-manager-565994456b-sh4t6\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.277016 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "25dfc3d0-6a18-48fd-9fcf-68313673e6b7" (UID: "25dfc3d0-6a18-48fd-9fcf-68313673e6b7"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.278282 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-client-ca" (OuterVolumeSpecName: "client-ca") pod "25dfc3d0-6a18-48fd-9fcf-68313673e6b7" (UID: "25dfc3d0-6a18-48fd-9fcf-68313673e6b7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.278426 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-config" (OuterVolumeSpecName: "config") pod "25dfc3d0-6a18-48fd-9fcf-68313673e6b7" (UID: "25dfc3d0-6a18-48fd-9fcf-68313673e6b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.278885 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-config" (OuterVolumeSpecName: "config") pod "a6b20452-fd7b-4169-8e38-d5405fa0cbb3" (UID: "a6b20452-fd7b-4169-8e38-d5405fa0cbb3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.279572 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-client-ca" (OuterVolumeSpecName: "client-ca") pod "a6b20452-fd7b-4169-8e38-d5405fa0cbb3" (UID: "a6b20452-fd7b-4169-8e38-d5405fa0cbb3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.283891 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-kube-api-access-chlpj" (OuterVolumeSpecName: "kube-api-access-chlpj") pod "a6b20452-fd7b-4169-8e38-d5405fa0cbb3" (UID: "a6b20452-fd7b-4169-8e38-d5405fa0cbb3"). InnerVolumeSpecName "kube-api-access-chlpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.284111 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "25dfc3d0-6a18-48fd-9fcf-68313673e6b7" (UID: "25dfc3d0-6a18-48fd-9fcf-68313673e6b7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.285519 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-kube-api-access-lrj5c" (OuterVolumeSpecName: "kube-api-access-lrj5c") pod "25dfc3d0-6a18-48fd-9fcf-68313673e6b7" (UID: "25dfc3d0-6a18-48fd-9fcf-68313673e6b7"). InnerVolumeSpecName "kube-api-access-lrj5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.286344 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a6b20452-fd7b-4169-8e38-d5405fa0cbb3" (UID: "a6b20452-fd7b-4169-8e38-d5405fa0cbb3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.377606 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-client-ca\") pod \"controller-manager-565994456b-sh4t6\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.377667 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-proxy-ca-bundles\") pod \"controller-manager-565994456b-sh4t6\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.377711 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-config\") pod \"controller-manager-565994456b-sh4t6\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.377748 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-serving-cert\") pod \"controller-manager-565994456b-sh4t6\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.377788 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kqn8\" (UniqueName: \"kubernetes.io/projected/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-kube-api-access-7kqn8\") pod \"controller-manager-565994456b-sh4t6\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.377887 4814 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.377901 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.377911 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chlpj\" (UniqueName: \"kubernetes.io/projected/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-kube-api-access-chlpj\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.377923 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.377932 4814 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.377941 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.377951 4814 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.377961 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrj5c\" (UniqueName: \"kubernetes.io/projected/25dfc3d0-6a18-48fd-9fcf-68313673e6b7-kube-api-access-lrj5c\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.377971 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6b20452-fd7b-4169-8e38-d5405fa0cbb3-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.378756 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-client-ca\") pod \"controller-manager-565994456b-sh4t6\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.379326 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-config\") pod \"controller-manager-565994456b-sh4t6\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.379479 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-proxy-ca-bundles\") pod \"controller-manager-565994456b-sh4t6\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.382283 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-serving-cert\") pod \"controller-manager-565994456b-sh4t6\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.397606 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kqn8\" (UniqueName: \"kubernetes.io/projected/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-kube-api-access-7kqn8\") pod \"controller-manager-565994456b-sh4t6\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.510748 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.605642 4814 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-07 05:08:18.83458767 +0000 UTC Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.605989 4814 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7524h40m32.228601752s for next certificate rotation Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.735112 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-565994456b-sh4t6"] Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.739124 4814 generic.go:334] "Generic (PLEG): container finished" podID="a6b20452-fd7b-4169-8e38-d5405fa0cbb3" containerID="e9eb5c9010c110dcd4eacb25fd7c74e465f10e847e8fd9ca9e8c4c8c0fb3b402" exitCode=0 Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.739363 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" event={"ID":"a6b20452-fd7b-4169-8e38-d5405fa0cbb3","Type":"ContainerDied","Data":"e9eb5c9010c110dcd4eacb25fd7c74e465f10e847e8fd9ca9e8c4c8c0fb3b402"} Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.739428 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" event={"ID":"a6b20452-fd7b-4169-8e38-d5405fa0cbb3","Type":"ContainerDied","Data":"5939f69cfd7643e39ea507709496007b5cfef7f9e1890a890f222fb71cd6222e"} Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.739473 4814 scope.go:117] "RemoveContainer" containerID="e9eb5c9010c110dcd4eacb25fd7c74e465f10e847e8fd9ca9e8c4c8c0fb3b402" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.739693 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.748575 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-972lm" event={"ID":"2cdfdb15-014d-40fc-9a7b-a905278f1249","Type":"ContainerStarted","Data":"3206866de8a504fead0c339b0b5d6119678de0b4b7a844b5b3ebf9a1ce36bead"} Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.752623 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ltqgw" event={"ID":"44a897a0-3e5d-4c66-b159-a8830705e7cf","Type":"ContainerStarted","Data":"7dc5095ffb44aee85d6f40bc3b201baafd1a78df6677f24782ea3f2d3f7de6fb"} Feb 27 16:27:46 crc kubenswrapper[4814]: W0227 16:27:46.756716 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ec92faf_a8c8_490f_95aa_caf7c3e00a2b.slice/crio-6259c3bb13bc35ad0d8d8c9dde16c8d243950febe10face485593554186b827e WatchSource:0}: Error finding container 6259c3bb13bc35ad0d8d8c9dde16c8d243950febe10face485593554186b827e: Status 404 returned error can't find the container with id 6259c3bb13bc35ad0d8d8c9dde16c8d243950febe10face485593554186b827e Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.756927 4814 generic.go:334] "Generic (PLEG): container finished" podID="8781d31c-4701-4e6b-8065-495f1e890c9f" containerID="f0d8a4412d8e779ac5926ee3f2fdf5680097423b4072d7ce9b4e0d49d5a87a8f" exitCode=0 Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.757003 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dj4vx" event={"ID":"8781d31c-4701-4e6b-8065-495f1e890c9f","Type":"ContainerDied","Data":"f0d8a4412d8e779ac5926ee3f2fdf5680097423b4072d7ce9b4e0d49d5a87a8f"} Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.761545 4814 generic.go:334] "Generic (PLEG): container finished" podID="9543a634-2567-4f4e-8e45-5c2eccb2344e" containerID="a9e458d9c2722ee4d535cc7b7040fff5ed2931f762d4313e4878afe2c9335015" exitCode=0 Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.761632 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9543a634-2567-4f4e-8e45-5c2eccb2344e","Type":"ContainerDied","Data":"a9e458d9c2722ee4d535cc7b7040fff5ed2931f762d4313e4878afe2c9335015"} Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.764900 4814 generic.go:334] "Generic (PLEG): container finished" podID="25dfc3d0-6a18-48fd-9fcf-68313673e6b7" containerID="2147b1f6d7df54d96f3b06111bef8a1dec30c4592cc9ec706bef48ab0ef1852e" exitCode=0 Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.765833 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.766234 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" event={"ID":"25dfc3d0-6a18-48fd-9fcf-68313673e6b7","Type":"ContainerDied","Data":"2147b1f6d7df54d96f3b06111bef8a1dec30c4592cc9ec706bef48ab0ef1852e"} Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.766559 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd" event={"ID":"25dfc3d0-6a18-48fd-9fcf-68313673e6b7","Type":"ContainerDied","Data":"59030b77db0b836e296f9e308578691b14c4c6dd9feec56eb747b4d25a56cd6b"} Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.768220 4814 scope.go:117] "RemoveContainer" containerID="e9eb5c9010c110dcd4eacb25fd7c74e465f10e847e8fd9ca9e8c4c8c0fb3b402" Feb 27 16:27:46 crc kubenswrapper[4814]: E0227 16:27:46.769311 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9eb5c9010c110dcd4eacb25fd7c74e465f10e847e8fd9ca9e8c4c8c0fb3b402\": container with ID starting with e9eb5c9010c110dcd4eacb25fd7c74e465f10e847e8fd9ca9e8c4c8c0fb3b402 not found: ID does not exist" containerID="e9eb5c9010c110dcd4eacb25fd7c74e465f10e847e8fd9ca9e8c4c8c0fb3b402" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.769339 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9eb5c9010c110dcd4eacb25fd7c74e465f10e847e8fd9ca9e8c4c8c0fb3b402"} err="failed to get container status \"e9eb5c9010c110dcd4eacb25fd7c74e465f10e847e8fd9ca9e8c4c8c0fb3b402\": rpc error: code = NotFound desc = could not find container \"e9eb5c9010c110dcd4eacb25fd7c74e465f10e847e8fd9ca9e8c4c8c0fb3b402\": container with ID starting with e9eb5c9010c110dcd4eacb25fd7c74e465f10e847e8fd9ca9e8c4c8c0fb3b402 not found: ID does not exist" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.769358 4814 scope.go:117] "RemoveContainer" containerID="2147b1f6d7df54d96f3b06111bef8a1dec30c4592cc9ec706bef48ab0ef1852e" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.784833 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-972lm" podStartSLOduration=2.593565044 podStartE2EDuration="47.784802606s" podCreationTimestamp="2026-02-27 16:26:59 +0000 UTC" firstStartedPulling="2026-02-27 16:27:01.219960034 +0000 UTC m=+233.672584864" lastFinishedPulling="2026-02-27 16:27:46.411197596 +0000 UTC m=+278.863822426" observedRunningTime="2026-02-27 16:27:46.774763773 +0000 UTC m=+279.227388623" watchObservedRunningTime="2026-02-27 16:27:46.784802606 +0000 UTC m=+279.237427436" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.796558 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd"] Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.800702 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6f9f8b7999-tmwhd"] Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.823558 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6"] Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.829774 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-869499dcbc-sq9h6"] Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.841106 4814 scope.go:117] "RemoveContainer" containerID="2147b1f6d7df54d96f3b06111bef8a1dec30c4592cc9ec706bef48ab0ef1852e" Feb 27 16:27:46 crc kubenswrapper[4814]: E0227 16:27:46.844101 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2147b1f6d7df54d96f3b06111bef8a1dec30c4592cc9ec706bef48ab0ef1852e\": container with ID starting with 2147b1f6d7df54d96f3b06111bef8a1dec30c4592cc9ec706bef48ab0ef1852e not found: ID does not exist" containerID="2147b1f6d7df54d96f3b06111bef8a1dec30c4592cc9ec706bef48ab0ef1852e" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.844131 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2147b1f6d7df54d96f3b06111bef8a1dec30c4592cc9ec706bef48ab0ef1852e"} err="failed to get container status \"2147b1f6d7df54d96f3b06111bef8a1dec30c4592cc9ec706bef48ab0ef1852e\": rpc error: code = NotFound desc = could not find container \"2147b1f6d7df54d96f3b06111bef8a1dec30c4592cc9ec706bef48ab0ef1852e\": container with ID starting with 2147b1f6d7df54d96f3b06111bef8a1dec30c4592cc9ec706bef48ab0ef1852e not found: ID does not exist" Feb 27 16:27:46 crc kubenswrapper[4814]: I0227 16:27:46.861102 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ltqgw" podStartSLOduration=9.821116116 podStartE2EDuration="45.861087061s" podCreationTimestamp="2026-02-27 16:27:01 +0000 UTC" firstStartedPulling="2026-02-27 16:27:10.05937265 +0000 UTC m=+242.511997510" lastFinishedPulling="2026-02-27 16:27:46.099343625 +0000 UTC m=+278.551968455" observedRunningTime="2026-02-27 16:27:46.859477636 +0000 UTC m=+279.312102476" watchObservedRunningTime="2026-02-27 16:27:46.861087061 +0000 UTC m=+279.313711891" Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.055116 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.138556 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536826-dgwn4" Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.190428 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/533e678e-abf5-45d8-bb78-db2f3d7b7928-kubelet-dir\") pod \"533e678e-abf5-45d8-bb78-db2f3d7b7928\" (UID: \"533e678e-abf5-45d8-bb78-db2f3d7b7928\") " Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.190568 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/533e678e-abf5-45d8-bb78-db2f3d7b7928-kube-api-access\") pod \"533e678e-abf5-45d8-bb78-db2f3d7b7928\" (UID: \"533e678e-abf5-45d8-bb78-db2f3d7b7928\") " Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.190518 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/533e678e-abf5-45d8-bb78-db2f3d7b7928-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "533e678e-abf5-45d8-bb78-db2f3d7b7928" (UID: "533e678e-abf5-45d8-bb78-db2f3d7b7928"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.190920 4814 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/533e678e-abf5-45d8-bb78-db2f3d7b7928-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.208454 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/533e678e-abf5-45d8-bb78-db2f3d7b7928-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "533e678e-abf5-45d8-bb78-db2f3d7b7928" (UID: "533e678e-abf5-45d8-bb78-db2f3d7b7928"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.291887 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w6zl\" (UniqueName: \"kubernetes.io/projected/95913cb4-39f1-44c0-ac49-0a2d51047679-kube-api-access-8w6zl\") pod \"95913cb4-39f1-44c0-ac49-0a2d51047679\" (UID: \"95913cb4-39f1-44c0-ac49-0a2d51047679\") " Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.292285 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/533e678e-abf5-45d8-bb78-db2f3d7b7928-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.297270 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95913cb4-39f1-44c0-ac49-0a2d51047679-kube-api-access-8w6zl" (OuterVolumeSpecName: "kube-api-access-8w6zl") pod "95913cb4-39f1-44c0-ac49-0a2d51047679" (UID: "95913cb4-39f1-44c0-ac49-0a2d51047679"). InnerVolumeSpecName "kube-api-access-8w6zl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.393315 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w6zl\" (UniqueName: \"kubernetes.io/projected/95913cb4-39f1-44c0-ac49-0a2d51047679-kube-api-access-8w6zl\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.606895 4814 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-15 06:22:59.700284649 +0000 UTC Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.608440 4814 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6253h55m12.091851024s for next certificate rotation Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.773922 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" event={"ID":"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b","Type":"ContainerStarted","Data":"a02afdd51ee070748e9b691122356254428d07f1f13a26a500446c428a0c0f2e"} Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.773975 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" event={"ID":"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b","Type":"ContainerStarted","Data":"6259c3bb13bc35ad0d8d8c9dde16c8d243950febe10face485593554186b827e"} Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.774281 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.776421 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536826-dgwn4" event={"ID":"95913cb4-39f1-44c0-ac49-0a2d51047679","Type":"ContainerDied","Data":"563fc3d178b9caef3700f56e12565007094ababa2341ca39792590befcbab569"} Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.776537 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="563fc3d178b9caef3700f56e12565007094ababa2341ca39792590befcbab569" Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.776656 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536826-dgwn4" Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.780151 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.780149 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"533e678e-abf5-45d8-bb78-db2f3d7b7928","Type":"ContainerDied","Data":"93b3b5312df6b856f944d69621378a67ba4337ae0faa8d52cd7f6fb4faa77ab9"} Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.781410 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93b3b5312df6b856f944d69621378a67ba4337ae0faa8d52cd7f6fb4faa77ab9" Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.790862 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.793566 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dj4vx" event={"ID":"8781d31c-4701-4e6b-8065-495f1e890c9f","Type":"ContainerStarted","Data":"4a18f5e10360fa49dbf1ade26714f62235c0a45decd458ac935f41395cf9d88e"} Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.803164 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" podStartSLOduration=12.803139625 podStartE2EDuration="12.803139625s" podCreationTimestamp="2026-02-27 16:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:27:47.799562572 +0000 UTC m=+280.252187402" watchObservedRunningTime="2026-02-27 16:27:47.803139625 +0000 UTC m=+280.255764455" Feb 27 16:27:47 crc kubenswrapper[4814]: I0227 16:27:47.829485 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dj4vx" podStartSLOduration=43.228903108 podStartE2EDuration="45.829460903s" podCreationTimestamp="2026-02-27 16:27:02 +0000 UTC" firstStartedPulling="2026-02-27 16:27:44.61670667 +0000 UTC m=+277.069331500" lastFinishedPulling="2026-02-27 16:27:47.217264465 +0000 UTC m=+279.669889295" observedRunningTime="2026-02-27 16:27:47.82906042 +0000 UTC m=+280.281685270" watchObservedRunningTime="2026-02-27 16:27:47.829460903 +0000 UTC m=+280.282085733" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.051004 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.203815 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9543a634-2567-4f4e-8e45-5c2eccb2344e-kube-api-access\") pod \"9543a634-2567-4f4e-8e45-5c2eccb2344e\" (UID: \"9543a634-2567-4f4e-8e45-5c2eccb2344e\") " Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.203887 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9543a634-2567-4f4e-8e45-5c2eccb2344e-kubelet-dir\") pod \"9543a634-2567-4f4e-8e45-5c2eccb2344e\" (UID: \"9543a634-2567-4f4e-8e45-5c2eccb2344e\") " Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.204011 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9543a634-2567-4f4e-8e45-5c2eccb2344e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9543a634-2567-4f4e-8e45-5c2eccb2344e" (UID: "9543a634-2567-4f4e-8e45-5c2eccb2344e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.204295 4814 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9543a634-2567-4f4e-8e45-5c2eccb2344e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.219058 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9543a634-2567-4f4e-8e45-5c2eccb2344e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9543a634-2567-4f4e-8e45-5c2eccb2344e" (UID: "9543a634-2567-4f4e-8e45-5c2eccb2344e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.306197 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9543a634-2567-4f4e-8e45-5c2eccb2344e-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.495009 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25dfc3d0-6a18-48fd-9fcf-68313673e6b7" path="/var/lib/kubelet/pods/25dfc3d0-6a18-48fd-9fcf-68313673e6b7/volumes" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.495814 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6b20452-fd7b-4169-8e38-d5405fa0cbb3" path="/var/lib/kubelet/pods/a6b20452-fd7b-4169-8e38-d5405fa0cbb3/volumes" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.521081 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx"] Feb 27 16:27:48 crc kubenswrapper[4814]: E0227 16:27:48.521397 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9543a634-2567-4f4e-8e45-5c2eccb2344e" containerName="pruner" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.521417 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="9543a634-2567-4f4e-8e45-5c2eccb2344e" containerName="pruner" Feb 27 16:27:48 crc kubenswrapper[4814]: E0227 16:27:48.521433 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="533e678e-abf5-45d8-bb78-db2f3d7b7928" containerName="pruner" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.521440 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="533e678e-abf5-45d8-bb78-db2f3d7b7928" containerName="pruner" Feb 27 16:27:48 crc kubenswrapper[4814]: E0227 16:27:48.521452 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95913cb4-39f1-44c0-ac49-0a2d51047679" containerName="oc" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.521458 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="95913cb4-39f1-44c0-ac49-0a2d51047679" containerName="oc" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.521552 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="9543a634-2567-4f4e-8e45-5c2eccb2344e" containerName="pruner" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.521563 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="533e678e-abf5-45d8-bb78-db2f3d7b7928" containerName="pruner" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.521578 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="95913cb4-39f1-44c0-ac49-0a2d51047679" containerName="oc" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.521963 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.542351 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx"] Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.542521 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.542897 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.543059 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.543224 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.543221 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.546239 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.711453 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0801bcb0-748f-4736-ad86-697aff3c5a10-config\") pod \"route-controller-manager-54f96c65c5-chjlx\" (UID: \"0801bcb0-748f-4736-ad86-697aff3c5a10\") " pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.711512 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr8tb\" (UniqueName: \"kubernetes.io/projected/0801bcb0-748f-4736-ad86-697aff3c5a10-kube-api-access-qr8tb\") pod \"route-controller-manager-54f96c65c5-chjlx\" (UID: \"0801bcb0-748f-4736-ad86-697aff3c5a10\") " pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.711764 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0801bcb0-748f-4736-ad86-697aff3c5a10-client-ca\") pod \"route-controller-manager-54f96c65c5-chjlx\" (UID: \"0801bcb0-748f-4736-ad86-697aff3c5a10\") " pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.711817 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0801bcb0-748f-4736-ad86-697aff3c5a10-serving-cert\") pod \"route-controller-manager-54f96c65c5-chjlx\" (UID: \"0801bcb0-748f-4736-ad86-697aff3c5a10\") " pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.813050 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0801bcb0-748f-4736-ad86-697aff3c5a10-client-ca\") pod \"route-controller-manager-54f96c65c5-chjlx\" (UID: \"0801bcb0-748f-4736-ad86-697aff3c5a10\") " pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.813094 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0801bcb0-748f-4736-ad86-697aff3c5a10-serving-cert\") pod \"route-controller-manager-54f96c65c5-chjlx\" (UID: \"0801bcb0-748f-4736-ad86-697aff3c5a10\") " pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.813175 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0801bcb0-748f-4736-ad86-697aff3c5a10-config\") pod \"route-controller-manager-54f96c65c5-chjlx\" (UID: \"0801bcb0-748f-4736-ad86-697aff3c5a10\") " pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.813202 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr8tb\" (UniqueName: \"kubernetes.io/projected/0801bcb0-748f-4736-ad86-697aff3c5a10-kube-api-access-qr8tb\") pod \"route-controller-manager-54f96c65c5-chjlx\" (UID: \"0801bcb0-748f-4736-ad86-697aff3c5a10\") " pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.814657 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0801bcb0-748f-4736-ad86-697aff3c5a10-config\") pod \"route-controller-manager-54f96c65c5-chjlx\" (UID: \"0801bcb0-748f-4736-ad86-697aff3c5a10\") " pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.814758 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9543a634-2567-4f4e-8e45-5c2eccb2344e","Type":"ContainerDied","Data":"161b234a09c9e0ed136413eddceaa8df01cc3511000699f6189891733ef0c00b"} Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.814796 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="161b234a09c9e0ed136413eddceaa8df01cc3511000699f6189891733ef0c00b" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.814875 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.816170 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0801bcb0-748f-4736-ad86-697aff3c5a10-client-ca\") pod \"route-controller-manager-54f96c65c5-chjlx\" (UID: \"0801bcb0-748f-4736-ad86-697aff3c5a10\") " pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.823285 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0801bcb0-748f-4736-ad86-697aff3c5a10-serving-cert\") pod \"route-controller-manager-54f96c65c5-chjlx\" (UID: \"0801bcb0-748f-4736-ad86-697aff3c5a10\") " pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.835779 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr8tb\" (UniqueName: \"kubernetes.io/projected/0801bcb0-748f-4736-ad86-697aff3c5a10-kube-api-access-qr8tb\") pod \"route-controller-manager-54f96c65c5-chjlx\" (UID: \"0801bcb0-748f-4736-ad86-697aff3c5a10\") " pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:27:48 crc kubenswrapper[4814]: I0227 16:27:48.838500 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:27:49 crc kubenswrapper[4814]: I0227 16:27:49.118206 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx"] Feb 27 16:27:49 crc kubenswrapper[4814]: W0227 16:27:49.124963 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0801bcb0_748f_4736_ad86_697aff3c5a10.slice/crio-4790a09a578f52cc4f4c55b1a49b04f58fa2d8ec15705564c09477137d256e8d WatchSource:0}: Error finding container 4790a09a578f52cc4f4c55b1a49b04f58fa2d8ec15705564c09477137d256e8d: Status 404 returned error can't find the container with id 4790a09a578f52cc4f4c55b1a49b04f58fa2d8ec15705564c09477137d256e8d Feb 27 16:27:49 crc kubenswrapper[4814]: I0227 16:27:49.820395 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-972lm" Feb 27 16:27:49 crc kubenswrapper[4814]: I0227 16:27:49.820975 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-972lm" Feb 27 16:27:49 crc kubenswrapper[4814]: I0227 16:27:49.825694 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" event={"ID":"0801bcb0-748f-4736-ad86-697aff3c5a10","Type":"ContainerStarted","Data":"e0128b6e8eb899bee164f8bb852738704d922a765b2f5e4bd8332715f13fecc3"} Feb 27 16:27:49 crc kubenswrapper[4814]: I0227 16:27:49.825921 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:27:49 crc kubenswrapper[4814]: I0227 16:27:49.826072 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" event={"ID":"0801bcb0-748f-4736-ad86-697aff3c5a10","Type":"ContainerStarted","Data":"4790a09a578f52cc4f4c55b1a49b04f58fa2d8ec15705564c09477137d256e8d"} Feb 27 16:27:49 crc kubenswrapper[4814]: I0227 16:27:49.839805 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:27:49 crc kubenswrapper[4814]: I0227 16:27:49.851635 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" podStartSLOduration=14.851594544 podStartE2EDuration="14.851594544s" podCreationTimestamp="2026-02-27 16:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:27:49.847606659 +0000 UTC m=+282.300231529" watchObservedRunningTime="2026-02-27 16:27:49.851594544 +0000 UTC m=+282.304219374" Feb 27 16:27:49 crc kubenswrapper[4814]: I0227 16:27:49.997679 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-972lm" Feb 27 16:27:51 crc kubenswrapper[4814]: I0227 16:27:51.629836 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:51 crc kubenswrapper[4814]: I0227 16:27:51.630286 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:51 crc kubenswrapper[4814]: I0227 16:27:51.678388 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:51 crc kubenswrapper[4814]: I0227 16:27:51.884514 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:52 crc kubenswrapper[4814]: I0227 16:27:52.850157 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:27:52 crc kubenswrapper[4814]: I0227 16:27:52.850712 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:27:52 crc kubenswrapper[4814]: I0227 16:27:52.902170 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:27:52 crc kubenswrapper[4814]: I0227 16:27:52.902244 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:27:53 crc kubenswrapper[4814]: I0227 16:27:53.899505 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dj4vx" podUID="8781d31c-4701-4e6b-8065-495f1e890c9f" containerName="registry-server" probeResult="failure" output=< Feb 27 16:27:53 crc kubenswrapper[4814]: timeout: failed to connect service ":50051" within 1s Feb 27 16:27:53 crc kubenswrapper[4814]: > Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.036212 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ltqgw"] Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.036575 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ltqgw" podUID="44a897a0-3e5d-4c66-b159-a8830705e7cf" containerName="registry-server" containerID="cri-o://7dc5095ffb44aee85d6f40bc3b201baafd1a78df6677f24782ea3f2d3f7de6fb" gracePeriod=2 Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.515200 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.621601 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44a897a0-3e5d-4c66-b159-a8830705e7cf-catalog-content\") pod \"44a897a0-3e5d-4c66-b159-a8830705e7cf\" (UID: \"44a897a0-3e5d-4c66-b159-a8830705e7cf\") " Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.621664 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44a897a0-3e5d-4c66-b159-a8830705e7cf-utilities\") pod \"44a897a0-3e5d-4c66-b159-a8830705e7cf\" (UID: \"44a897a0-3e5d-4c66-b159-a8830705e7cf\") " Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.621751 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgs7w\" (UniqueName: \"kubernetes.io/projected/44a897a0-3e5d-4c66-b159-a8830705e7cf-kube-api-access-hgs7w\") pod \"44a897a0-3e5d-4c66-b159-a8830705e7cf\" (UID: \"44a897a0-3e5d-4c66-b159-a8830705e7cf\") " Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.625266 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44a897a0-3e5d-4c66-b159-a8830705e7cf-utilities" (OuterVolumeSpecName: "utilities") pod "44a897a0-3e5d-4c66-b159-a8830705e7cf" (UID: "44a897a0-3e5d-4c66-b159-a8830705e7cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.628506 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44a897a0-3e5d-4c66-b159-a8830705e7cf-kube-api-access-hgs7w" (OuterVolumeSpecName: "kube-api-access-hgs7w") pod "44a897a0-3e5d-4c66-b159-a8830705e7cf" (UID: "44a897a0-3e5d-4c66-b159-a8830705e7cf"). InnerVolumeSpecName "kube-api-access-hgs7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.651496 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44a897a0-3e5d-4c66-b159-a8830705e7cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44a897a0-3e5d-4c66-b159-a8830705e7cf" (UID: "44a897a0-3e5d-4c66-b159-a8830705e7cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.723426 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44a897a0-3e5d-4c66-b159-a8830705e7cf-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.724033 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44a897a0-3e5d-4c66-b159-a8830705e7cf-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.724102 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgs7w\" (UniqueName: \"kubernetes.io/projected/44a897a0-3e5d-4c66-b159-a8830705e7cf-kube-api-access-hgs7w\") on node \"crc\" DevicePath \"\"" Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.855761 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cm8p5" event={"ID":"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375","Type":"ContainerStarted","Data":"208fbbd20e9ec61739404ddb25a8913c15f9d97180bddccb121fd850e60545c0"} Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.858075 4814 generic.go:334] "Generic (PLEG): container finished" podID="44a897a0-3e5d-4c66-b159-a8830705e7cf" containerID="7dc5095ffb44aee85d6f40bc3b201baafd1a78df6677f24782ea3f2d3f7de6fb" exitCode=0 Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.858137 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ltqgw" event={"ID":"44a897a0-3e5d-4c66-b159-a8830705e7cf","Type":"ContainerDied","Data":"7dc5095ffb44aee85d6f40bc3b201baafd1a78df6677f24782ea3f2d3f7de6fb"} Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.858156 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ltqgw" event={"ID":"44a897a0-3e5d-4c66-b159-a8830705e7cf","Type":"ContainerDied","Data":"fd8f35f8dd59b93a9ad665b2251c172c40a2a828877279d623c095f7d218a603"} Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.858175 4814 scope.go:117] "RemoveContainer" containerID="7dc5095ffb44aee85d6f40bc3b201baafd1a78df6677f24782ea3f2d3f7de6fb" Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.858341 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ltqgw" Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.861574 4814 generic.go:334] "Generic (PLEG): container finished" podID="5df9c141-ab4f-40e3-b733-da2a2712a326" containerID="74ba1ee457e00edd1799583460cb1f1d4bae89ff1fb268ce58bff562a9ff515f" exitCode=0 Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.861619 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zbl6" event={"ID":"5df9c141-ab4f-40e3-b733-da2a2712a326","Type":"ContainerDied","Data":"74ba1ee457e00edd1799583460cb1f1d4bae89ff1fb268ce58bff562a9ff515f"} Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.882602 4814 scope.go:117] "RemoveContainer" containerID="bfe8a1cc35dcf3fd440b085311f857ee97d33222ba4804011c9850b7554e4f3e" Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.917179 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ltqgw"] Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.919987 4814 scope.go:117] "RemoveContainer" containerID="11bac898d6d6afd051affa139d42a1071f6277d1c8310efe89949a7187f1a9ce" Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.921942 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ltqgw"] Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.940415 4814 scope.go:117] "RemoveContainer" containerID="7dc5095ffb44aee85d6f40bc3b201baafd1a78df6677f24782ea3f2d3f7de6fb" Feb 27 16:27:54 crc kubenswrapper[4814]: E0227 16:27:54.941467 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dc5095ffb44aee85d6f40bc3b201baafd1a78df6677f24782ea3f2d3f7de6fb\": container with ID starting with 7dc5095ffb44aee85d6f40bc3b201baafd1a78df6677f24782ea3f2d3f7de6fb not found: ID does not exist" containerID="7dc5095ffb44aee85d6f40bc3b201baafd1a78df6677f24782ea3f2d3f7de6fb" Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.941530 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dc5095ffb44aee85d6f40bc3b201baafd1a78df6677f24782ea3f2d3f7de6fb"} err="failed to get container status \"7dc5095ffb44aee85d6f40bc3b201baafd1a78df6677f24782ea3f2d3f7de6fb\": rpc error: code = NotFound desc = could not find container \"7dc5095ffb44aee85d6f40bc3b201baafd1a78df6677f24782ea3f2d3f7de6fb\": container with ID starting with 7dc5095ffb44aee85d6f40bc3b201baafd1a78df6677f24782ea3f2d3f7de6fb not found: ID does not exist" Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.941568 4814 scope.go:117] "RemoveContainer" containerID="bfe8a1cc35dcf3fd440b085311f857ee97d33222ba4804011c9850b7554e4f3e" Feb 27 16:27:54 crc kubenswrapper[4814]: E0227 16:27:54.942421 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfe8a1cc35dcf3fd440b085311f857ee97d33222ba4804011c9850b7554e4f3e\": container with ID starting with bfe8a1cc35dcf3fd440b085311f857ee97d33222ba4804011c9850b7554e4f3e not found: ID does not exist" containerID="bfe8a1cc35dcf3fd440b085311f857ee97d33222ba4804011c9850b7554e4f3e" Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.942477 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfe8a1cc35dcf3fd440b085311f857ee97d33222ba4804011c9850b7554e4f3e"} err="failed to get container status \"bfe8a1cc35dcf3fd440b085311f857ee97d33222ba4804011c9850b7554e4f3e\": rpc error: code = NotFound desc = could not find container \"bfe8a1cc35dcf3fd440b085311f857ee97d33222ba4804011c9850b7554e4f3e\": container with ID starting with bfe8a1cc35dcf3fd440b085311f857ee97d33222ba4804011c9850b7554e4f3e not found: ID does not exist" Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.942509 4814 scope.go:117] "RemoveContainer" containerID="11bac898d6d6afd051affa139d42a1071f6277d1c8310efe89949a7187f1a9ce" Feb 27 16:27:54 crc kubenswrapper[4814]: E0227 16:27:54.942887 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11bac898d6d6afd051affa139d42a1071f6277d1c8310efe89949a7187f1a9ce\": container with ID starting with 11bac898d6d6afd051affa139d42a1071f6277d1c8310efe89949a7187f1a9ce not found: ID does not exist" containerID="11bac898d6d6afd051affa139d42a1071f6277d1c8310efe89949a7187f1a9ce" Feb 27 16:27:54 crc kubenswrapper[4814]: I0227 16:27:54.942918 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11bac898d6d6afd051affa139d42a1071f6277d1c8310efe89949a7187f1a9ce"} err="failed to get container status \"11bac898d6d6afd051affa139d42a1071f6277d1c8310efe89949a7187f1a9ce\": rpc error: code = NotFound desc = could not find container \"11bac898d6d6afd051affa139d42a1071f6277d1c8310efe89949a7187f1a9ce\": container with ID starting with 11bac898d6d6afd051affa139d42a1071f6277d1c8310efe89949a7187f1a9ce not found: ID does not exist" Feb 27 16:27:55 crc kubenswrapper[4814]: I0227 16:27:55.870938 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zbl6" event={"ID":"5df9c141-ab4f-40e3-b733-da2a2712a326","Type":"ContainerStarted","Data":"81e8fe574c70824a561cd7739c591fe5f5ec131c54b59a261eb90b69a3f973d0"} Feb 27 16:27:55 crc kubenswrapper[4814]: I0227 16:27:55.872989 4814 generic.go:334] "Generic (PLEG): container finished" podID="f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" containerID="208fbbd20e9ec61739404ddb25a8913c15f9d97180bddccb121fd850e60545c0" exitCode=0 Feb 27 16:27:55 crc kubenswrapper[4814]: I0227 16:27:55.873031 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cm8p5" event={"ID":"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375","Type":"ContainerDied","Data":"208fbbd20e9ec61739404ddb25a8913c15f9d97180bddccb121fd850e60545c0"} Feb 27 16:27:55 crc kubenswrapper[4814]: I0227 16:27:55.922662 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4zbl6" podStartSLOduration=10.565706773 podStartE2EDuration="55.922641376s" podCreationTimestamp="2026-02-27 16:27:00 +0000 UTC" firstStartedPulling="2026-02-27 16:27:10.059015048 +0000 UTC m=+242.511639878" lastFinishedPulling="2026-02-27 16:27:55.415949651 +0000 UTC m=+287.868574481" observedRunningTime="2026-02-27 16:27:55.894528376 +0000 UTC m=+288.347153216" watchObservedRunningTime="2026-02-27 16:27:55.922641376 +0000 UTC m=+288.375266226" Feb 27 16:27:56 crc kubenswrapper[4814]: I0227 16:27:56.496295 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44a897a0-3e5d-4c66-b159-a8830705e7cf" path="/var/lib/kubelet/pods/44a897a0-3e5d-4c66-b159-a8830705e7cf/volumes" Feb 27 16:27:57 crc kubenswrapper[4814]: I0227 16:27:57.888474 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cm8p5" event={"ID":"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375","Type":"ContainerStarted","Data":"87931504b222c833ffc5fd5cb8dc55a426269355d7d33123a4f5c0c883e15d3a"} Feb 27 16:27:57 crc kubenswrapper[4814]: I0227 16:27:57.916734 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cm8p5" podStartSLOduration=3.950494977 podStartE2EDuration="59.916712738s" podCreationTimestamp="2026-02-27 16:26:58 +0000 UTC" firstStartedPulling="2026-02-27 16:27:01.229707941 +0000 UTC m=+233.682332771" lastFinishedPulling="2026-02-27 16:27:57.195925702 +0000 UTC m=+289.648550532" observedRunningTime="2026-02-27 16:27:57.913828389 +0000 UTC m=+290.366453229" watchObservedRunningTime="2026-02-27 16:27:57.916712738 +0000 UTC m=+290.369337578" Feb 27 16:27:58 crc kubenswrapper[4814]: I0227 16:27:58.896366 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ftgk" event={"ID":"b80cc2cd-d07d-4225-b078-9c70d8e189c7","Type":"ContainerStarted","Data":"c93f06679f23c4e6585f42ea61bdf308945fe55fac7aeb6a0b8f3afe9c640c8f"} Feb 27 16:27:59 crc kubenswrapper[4814]: I0227 16:27:59.244196 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:27:59 crc kubenswrapper[4814]: I0227 16:27:59.244306 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:27:59 crc kubenswrapper[4814]: I0227 16:27:59.863548 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-972lm" Feb 27 16:27:59 crc kubenswrapper[4814]: I0227 16:27:59.908096 4814 generic.go:334] "Generic (PLEG): container finished" podID="b80cc2cd-d07d-4225-b078-9c70d8e189c7" containerID="c93f06679f23c4e6585f42ea61bdf308945fe55fac7aeb6a0b8f3afe9c640c8f" exitCode=0 Feb 27 16:27:59 crc kubenswrapper[4814]: I0227 16:27:59.908195 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ftgk" event={"ID":"b80cc2cd-d07d-4225-b078-9c70d8e189c7","Type":"ContainerDied","Data":"c93f06679f23c4e6585f42ea61bdf308945fe55fac7aeb6a0b8f3afe9c640c8f"} Feb 27 16:27:59 crc kubenswrapper[4814]: I0227 16:27:59.911407 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sk6rs" event={"ID":"41b851a0-9bda-4e20-a692-5f796f55a749","Type":"ContainerStarted","Data":"07575168eaf9cbb5d7cc834dc2a531a5cb850e0efe81d9f5fffeb1286239e917"} Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.140581 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536828-g5rhp"] Feb 27 16:28:00 crc kubenswrapper[4814]: E0227 16:28:00.140912 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44a897a0-3e5d-4c66-b159-a8830705e7cf" containerName="registry-server" Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.140932 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="44a897a0-3e5d-4c66-b159-a8830705e7cf" containerName="registry-server" Feb 27 16:28:00 crc kubenswrapper[4814]: E0227 16:28:00.140945 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44a897a0-3e5d-4c66-b159-a8830705e7cf" containerName="extract-utilities" Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.140953 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="44a897a0-3e5d-4c66-b159-a8830705e7cf" containerName="extract-utilities" Feb 27 16:28:00 crc kubenswrapper[4814]: E0227 16:28:00.140963 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44a897a0-3e5d-4c66-b159-a8830705e7cf" containerName="extract-content" Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.140978 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="44a897a0-3e5d-4c66-b159-a8830705e7cf" containerName="extract-content" Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.141135 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="44a897a0-3e5d-4c66-b159-a8830705e7cf" containerName="registry-server" Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.141704 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536828-g5rhp" Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.144448 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.144683 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.145314 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.150245 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536828-g5rhp"] Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.289290 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-cm8p5" podUID="f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" containerName="registry-server" probeResult="failure" output=< Feb 27 16:28:00 crc kubenswrapper[4814]: timeout: failed to connect service ":50051" within 1s Feb 27 16:28:00 crc kubenswrapper[4814]: > Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.344582 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg25x\" (UniqueName: \"kubernetes.io/projected/dd3bb66c-e36b-456f-8051-509099d63405-kube-api-access-cg25x\") pod \"auto-csr-approver-29536828-g5rhp\" (UID: \"dd3bb66c-e36b-456f-8051-509099d63405\") " pod="openshift-infra/auto-csr-approver-29536828-g5rhp" Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.446835 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg25x\" (UniqueName: \"kubernetes.io/projected/dd3bb66c-e36b-456f-8051-509099d63405-kube-api-access-cg25x\") pod \"auto-csr-approver-29536828-g5rhp\" (UID: \"dd3bb66c-e36b-456f-8051-509099d63405\") " pod="openshift-infra/auto-csr-approver-29536828-g5rhp" Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.467657 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg25x\" (UniqueName: \"kubernetes.io/projected/dd3bb66c-e36b-456f-8051-509099d63405-kube-api-access-cg25x\") pod \"auto-csr-approver-29536828-g5rhp\" (UID: \"dd3bb66c-e36b-456f-8051-509099d63405\") " pod="openshift-infra/auto-csr-approver-29536828-g5rhp" Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.492687 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536828-g5rhp" Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.919575 4814 generic.go:334] "Generic (PLEG): container finished" podID="41b851a0-9bda-4e20-a692-5f796f55a749" containerID="07575168eaf9cbb5d7cc834dc2a531a5cb850e0efe81d9f5fffeb1286239e917" exitCode=0 Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.919973 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sk6rs" event={"ID":"41b851a0-9bda-4e20-a692-5f796f55a749","Type":"ContainerDied","Data":"07575168eaf9cbb5d7cc834dc2a531a5cb850e0efe81d9f5fffeb1286239e917"} Feb 27 16:28:00 crc kubenswrapper[4814]: I0227 16:28:00.943222 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536828-g5rhp"] Feb 27 16:28:01 crc kubenswrapper[4814]: I0227 16:28:01.199951 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:28:01 crc kubenswrapper[4814]: I0227 16:28:01.200039 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:28:01 crc kubenswrapper[4814]: I0227 16:28:01.257882 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:28:01 crc kubenswrapper[4814]: I0227 16:28:01.838827 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-972lm"] Feb 27 16:28:01 crc kubenswrapper[4814]: I0227 16:28:01.839063 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-972lm" podUID="2cdfdb15-014d-40fc-9a7b-a905278f1249" containerName="registry-server" containerID="cri-o://3206866de8a504fead0c339b0b5d6119678de0b4b7a844b5b3ebf9a1ce36bead" gracePeriod=2 Feb 27 16:28:01 crc kubenswrapper[4814]: I0227 16:28:01.930845 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536828-g5rhp" event={"ID":"dd3bb66c-e36b-456f-8051-509099d63405","Type":"ContainerStarted","Data":"bc91ba7d801bdcb10fd331735b5ffffecb3150f9b3ab387e8a0358818bb4cf99"} Feb 27 16:28:01 crc kubenswrapper[4814]: I0227 16:28:01.986440 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:28:02 crc kubenswrapper[4814]: I0227 16:28:02.889623 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:28:02 crc kubenswrapper[4814]: I0227 16:28:02.949100 4814 generic.go:334] "Generic (PLEG): container finished" podID="2cdfdb15-014d-40fc-9a7b-a905278f1249" containerID="3206866de8a504fead0c339b0b5d6119678de0b4b7a844b5b3ebf9a1ce36bead" exitCode=0 Feb 27 16:28:02 crc kubenswrapper[4814]: I0227 16:28:02.949870 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-972lm" event={"ID":"2cdfdb15-014d-40fc-9a7b-a905278f1249","Type":"ContainerDied","Data":"3206866de8a504fead0c339b0b5d6119678de0b4b7a844b5b3ebf9a1ce36bead"} Feb 27 16:28:02 crc kubenswrapper[4814]: I0227 16:28:02.956291 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:28:03 crc kubenswrapper[4814]: I0227 16:28:03.530922 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-972lm" Feb 27 16:28:03 crc kubenswrapper[4814]: I0227 16:28:03.697380 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmsmv\" (UniqueName: \"kubernetes.io/projected/2cdfdb15-014d-40fc-9a7b-a905278f1249-kube-api-access-pmsmv\") pod \"2cdfdb15-014d-40fc-9a7b-a905278f1249\" (UID: \"2cdfdb15-014d-40fc-9a7b-a905278f1249\") " Feb 27 16:28:03 crc kubenswrapper[4814]: I0227 16:28:03.697578 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cdfdb15-014d-40fc-9a7b-a905278f1249-catalog-content\") pod \"2cdfdb15-014d-40fc-9a7b-a905278f1249\" (UID: \"2cdfdb15-014d-40fc-9a7b-a905278f1249\") " Feb 27 16:28:03 crc kubenswrapper[4814]: I0227 16:28:03.697608 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cdfdb15-014d-40fc-9a7b-a905278f1249-utilities\") pod \"2cdfdb15-014d-40fc-9a7b-a905278f1249\" (UID: \"2cdfdb15-014d-40fc-9a7b-a905278f1249\") " Feb 27 16:28:03 crc kubenswrapper[4814]: I0227 16:28:03.698762 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cdfdb15-014d-40fc-9a7b-a905278f1249-utilities" (OuterVolumeSpecName: "utilities") pod "2cdfdb15-014d-40fc-9a7b-a905278f1249" (UID: "2cdfdb15-014d-40fc-9a7b-a905278f1249"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:28:03 crc kubenswrapper[4814]: I0227 16:28:03.708759 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cdfdb15-014d-40fc-9a7b-a905278f1249-kube-api-access-pmsmv" (OuterVolumeSpecName: "kube-api-access-pmsmv") pod "2cdfdb15-014d-40fc-9a7b-a905278f1249" (UID: "2cdfdb15-014d-40fc-9a7b-a905278f1249"). InnerVolumeSpecName "kube-api-access-pmsmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:28:03 crc kubenswrapper[4814]: I0227 16:28:03.763152 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cdfdb15-014d-40fc-9a7b-a905278f1249-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2cdfdb15-014d-40fc-9a7b-a905278f1249" (UID: "2cdfdb15-014d-40fc-9a7b-a905278f1249"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:28:03 crc kubenswrapper[4814]: I0227 16:28:03.800112 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cdfdb15-014d-40fc-9a7b-a905278f1249-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:03 crc kubenswrapper[4814]: I0227 16:28:03.800164 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cdfdb15-014d-40fc-9a7b-a905278f1249-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:03 crc kubenswrapper[4814]: I0227 16:28:03.800184 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmsmv\" (UniqueName: \"kubernetes.io/projected/2cdfdb15-014d-40fc-9a7b-a905278f1249-kube-api-access-pmsmv\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:03 crc kubenswrapper[4814]: I0227 16:28:03.958026 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-972lm" event={"ID":"2cdfdb15-014d-40fc-9a7b-a905278f1249","Type":"ContainerDied","Data":"113a0bc7b7414845c6dfa565ec8e4b4b54dd39570fee9bed75caeecf391e390d"} Feb 27 16:28:03 crc kubenswrapper[4814]: I0227 16:28:03.958113 4814 scope.go:117] "RemoveContainer" containerID="3206866de8a504fead0c339b0b5d6119678de0b4b7a844b5b3ebf9a1ce36bead" Feb 27 16:28:03 crc kubenswrapper[4814]: I0227 16:28:03.958126 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-972lm" Feb 27 16:28:04 crc kubenswrapper[4814]: I0227 16:28:04.001158 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-972lm"] Feb 27 16:28:04 crc kubenswrapper[4814]: I0227 16:28:04.005476 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-972lm"] Feb 27 16:28:04 crc kubenswrapper[4814]: I0227 16:28:04.499650 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cdfdb15-014d-40fc-9a7b-a905278f1249" path="/var/lib/kubelet/pods/2cdfdb15-014d-40fc-9a7b-a905278f1249/volumes" Feb 27 16:28:04 crc kubenswrapper[4814]: I0227 16:28:04.662004 4814 scope.go:117] "RemoveContainer" containerID="eecc9514e37f521f316d91ee6da545090202782ae23267acbcd118ea880c0edd" Feb 27 16:28:04 crc kubenswrapper[4814]: I0227 16:28:04.760712 4814 scope.go:117] "RemoveContainer" containerID="d49a517eb4c3d0ee712ade3917d8b9c94bcf5a6260f8f0f00ccda868f1d3c25c" Feb 27 16:28:06 crc kubenswrapper[4814]: I0227 16:28:06.432709 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dj4vx"] Feb 27 16:28:06 crc kubenswrapper[4814]: I0227 16:28:06.433017 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dj4vx" podUID="8781d31c-4701-4e6b-8065-495f1e890c9f" containerName="registry-server" containerID="cri-o://4a18f5e10360fa49dbf1ade26714f62235c0a45decd458ac935f41395cf9d88e" gracePeriod=2 Feb 27 16:28:07 crc kubenswrapper[4814]: I0227 16:28:07.828781 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:28:07 crc kubenswrapper[4814]: I0227 16:28:07.973868 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn5b7\" (UniqueName: \"kubernetes.io/projected/8781d31c-4701-4e6b-8065-495f1e890c9f-kube-api-access-zn5b7\") pod \"8781d31c-4701-4e6b-8065-495f1e890c9f\" (UID: \"8781d31c-4701-4e6b-8065-495f1e890c9f\") " Feb 27 16:28:07 crc kubenswrapper[4814]: I0227 16:28:07.973966 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8781d31c-4701-4e6b-8065-495f1e890c9f-utilities\") pod \"8781d31c-4701-4e6b-8065-495f1e890c9f\" (UID: \"8781d31c-4701-4e6b-8065-495f1e890c9f\") " Feb 27 16:28:07 crc kubenswrapper[4814]: I0227 16:28:07.974000 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8781d31c-4701-4e6b-8065-495f1e890c9f-catalog-content\") pod \"8781d31c-4701-4e6b-8065-495f1e890c9f\" (UID: \"8781d31c-4701-4e6b-8065-495f1e890c9f\") " Feb 27 16:28:07 crc kubenswrapper[4814]: I0227 16:28:07.974996 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8781d31c-4701-4e6b-8065-495f1e890c9f-utilities" (OuterVolumeSpecName: "utilities") pod "8781d31c-4701-4e6b-8065-495f1e890c9f" (UID: "8781d31c-4701-4e6b-8065-495f1e890c9f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:28:07 crc kubenswrapper[4814]: I0227 16:28:07.979361 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8781d31c-4701-4e6b-8065-495f1e890c9f-kube-api-access-zn5b7" (OuterVolumeSpecName: "kube-api-access-zn5b7") pod "8781d31c-4701-4e6b-8065-495f1e890c9f" (UID: "8781d31c-4701-4e6b-8065-495f1e890c9f"). InnerVolumeSpecName "kube-api-access-zn5b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.002623 4814 generic.go:334] "Generic (PLEG): container finished" podID="8781d31c-4701-4e6b-8065-495f1e890c9f" containerID="4a18f5e10360fa49dbf1ade26714f62235c0a45decd458ac935f41395cf9d88e" exitCode=0 Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.002740 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dj4vx" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.002745 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dj4vx" event={"ID":"8781d31c-4701-4e6b-8065-495f1e890c9f","Type":"ContainerDied","Data":"4a18f5e10360fa49dbf1ade26714f62235c0a45decd458ac935f41395cf9d88e"} Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.002841 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dj4vx" event={"ID":"8781d31c-4701-4e6b-8065-495f1e890c9f","Type":"ContainerDied","Data":"35ad2f22f8edc19859c225299b6e91f69a841b34b4031f20efa04b1ab40e607c"} Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.002868 4814 scope.go:117] "RemoveContainer" containerID="4a18f5e10360fa49dbf1ade26714f62235c0a45decd458ac935f41395cf9d88e" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.006800 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ftgk" event={"ID":"b80cc2cd-d07d-4225-b078-9c70d8e189c7","Type":"ContainerStarted","Data":"42074e214640646aa655a689cfe857d210987bcec894e98a7444ec78886ef8ec"} Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.008904 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536828-g5rhp" event={"ID":"dd3bb66c-e36b-456f-8051-509099d63405","Type":"ContainerStarted","Data":"81ef0dcfb1aa0d2d1bba5a21ad84b96dea5ed110e08c21309ab4e1e86a512043"} Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.015420 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sk6rs" event={"ID":"41b851a0-9bda-4e20-a692-5f796f55a749","Type":"ContainerStarted","Data":"368c23bbb8809ac49a66f06ee8351d5dab07d09c3f389c98854f5e3a6b21c5dd"} Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.017635 4814 generic.go:334] "Generic (PLEG): container finished" podID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" containerID="1899b6570d51daf80aa21a3f009ab53315427b24f79c18be45988a8e74237c7e" exitCode=0 Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.017679 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c5wns" event={"ID":"d3a647cf-eb96-40b8-982c-8e8b869e57c4","Type":"ContainerDied","Data":"1899b6570d51daf80aa21a3f009ab53315427b24f79c18be45988a8e74237c7e"} Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.022969 4814 scope.go:117] "RemoveContainer" containerID="f0d8a4412d8e779ac5926ee3f2fdf5680097423b4072d7ce9b4e0d49d5a87a8f" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.025897 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9ftgk" podStartSLOduration=16.896463511 podStartE2EDuration="1m6.02588756s" podCreationTimestamp="2026-02-27 16:27:02 +0000 UTC" firstStartedPulling="2026-02-27 16:27:15.532156752 +0000 UTC m=+247.984781582" lastFinishedPulling="2026-02-27 16:28:04.661580761 +0000 UTC m=+297.114205631" observedRunningTime="2026-02-27 16:28:08.024179582 +0000 UTC m=+300.476804422" watchObservedRunningTime="2026-02-27 16:28:08.02588756 +0000 UTC m=+300.478512390" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.049918 4814 scope.go:117] "RemoveContainer" containerID="dc5ad3d7ab001b79bec729342ae0f4e69e413119cb592d087f67b00d0eae5d68" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.068505 4814 scope.go:117] "RemoveContainer" containerID="4a18f5e10360fa49dbf1ade26714f62235c0a45decd458ac935f41395cf9d88e" Feb 27 16:28:08 crc kubenswrapper[4814]: E0227 16:28:08.068972 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a18f5e10360fa49dbf1ade26714f62235c0a45decd458ac935f41395cf9d88e\": container with ID starting with 4a18f5e10360fa49dbf1ade26714f62235c0a45decd458ac935f41395cf9d88e not found: ID does not exist" containerID="4a18f5e10360fa49dbf1ade26714f62235c0a45decd458ac935f41395cf9d88e" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.069015 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a18f5e10360fa49dbf1ade26714f62235c0a45decd458ac935f41395cf9d88e"} err="failed to get container status \"4a18f5e10360fa49dbf1ade26714f62235c0a45decd458ac935f41395cf9d88e\": rpc error: code = NotFound desc = could not find container \"4a18f5e10360fa49dbf1ade26714f62235c0a45decd458ac935f41395cf9d88e\": container with ID starting with 4a18f5e10360fa49dbf1ade26714f62235c0a45decd458ac935f41395cf9d88e not found: ID does not exist" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.069037 4814 scope.go:117] "RemoveContainer" containerID="f0d8a4412d8e779ac5926ee3f2fdf5680097423b4072d7ce9b4e0d49d5a87a8f" Feb 27 16:28:08 crc kubenswrapper[4814]: E0227 16:28:08.069385 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0d8a4412d8e779ac5926ee3f2fdf5680097423b4072d7ce9b4e0d49d5a87a8f\": container with ID starting with f0d8a4412d8e779ac5926ee3f2fdf5680097423b4072d7ce9b4e0d49d5a87a8f not found: ID does not exist" containerID="f0d8a4412d8e779ac5926ee3f2fdf5680097423b4072d7ce9b4e0d49d5a87a8f" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.069409 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0d8a4412d8e779ac5926ee3f2fdf5680097423b4072d7ce9b4e0d49d5a87a8f"} err="failed to get container status \"f0d8a4412d8e779ac5926ee3f2fdf5680097423b4072d7ce9b4e0d49d5a87a8f\": rpc error: code = NotFound desc = could not find container \"f0d8a4412d8e779ac5926ee3f2fdf5680097423b4072d7ce9b4e0d49d5a87a8f\": container with ID starting with f0d8a4412d8e779ac5926ee3f2fdf5680097423b4072d7ce9b4e0d49d5a87a8f not found: ID does not exist" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.069424 4814 scope.go:117] "RemoveContainer" containerID="dc5ad3d7ab001b79bec729342ae0f4e69e413119cb592d087f67b00d0eae5d68" Feb 27 16:28:08 crc kubenswrapper[4814]: E0227 16:28:08.069634 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc5ad3d7ab001b79bec729342ae0f4e69e413119cb592d087f67b00d0eae5d68\": container with ID starting with dc5ad3d7ab001b79bec729342ae0f4e69e413119cb592d087f67b00d0eae5d68 not found: ID does not exist" containerID="dc5ad3d7ab001b79bec729342ae0f4e69e413119cb592d087f67b00d0eae5d68" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.069655 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc5ad3d7ab001b79bec729342ae0f4e69e413119cb592d087f67b00d0eae5d68"} err="failed to get container status \"dc5ad3d7ab001b79bec729342ae0f4e69e413119cb592d087f67b00d0eae5d68\": rpc error: code = NotFound desc = could not find container \"dc5ad3d7ab001b79bec729342ae0f4e69e413119cb592d087f67b00d0eae5d68\": container with ID starting with dc5ad3d7ab001b79bec729342ae0f4e69e413119cb592d087f67b00d0eae5d68 not found: ID does not exist" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.076306 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn5b7\" (UniqueName: \"kubernetes.io/projected/8781d31c-4701-4e6b-8065-495f1e890c9f-kube-api-access-zn5b7\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.076327 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8781d31c-4701-4e6b-8065-495f1e890c9f-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.081159 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sk6rs" podStartSLOduration=3.046947085 podStartE2EDuration="1m9.081138968s" podCreationTimestamp="2026-02-27 16:26:59 +0000 UTC" firstStartedPulling="2026-02-27 16:27:01.213460129 +0000 UTC m=+233.666084959" lastFinishedPulling="2026-02-27 16:28:07.247652012 +0000 UTC m=+299.700276842" observedRunningTime="2026-02-27 16:28:08.05926665 +0000 UTC m=+300.511891480" watchObservedRunningTime="2026-02-27 16:28:08.081138968 +0000 UTC m=+300.533763798" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.081690 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536828-g5rhp" podStartSLOduration=1.794750112 podStartE2EDuration="8.081686346s" podCreationTimestamp="2026-02-27 16:28:00 +0000 UTC" firstStartedPulling="2026-02-27 16:28:00.959572159 +0000 UTC m=+293.412196989" lastFinishedPulling="2026-02-27 16:28:07.246508393 +0000 UTC m=+299.699133223" observedRunningTime="2026-02-27 16:28:08.078165916 +0000 UTC m=+300.530790746" watchObservedRunningTime="2026-02-27 16:28:08.081686346 +0000 UTC m=+300.534311176" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.129224 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8781d31c-4701-4e6b-8065-495f1e890c9f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8781d31c-4701-4e6b-8065-495f1e890c9f" (UID: "8781d31c-4701-4e6b-8065-495f1e890c9f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.177336 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8781d31c-4701-4e6b-8065-495f1e890c9f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.333284 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dj4vx"] Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.338452 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dj4vx"] Feb 27 16:28:08 crc kubenswrapper[4814]: I0227 16:28:08.499273 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8781d31c-4701-4e6b-8065-495f1e890c9f" path="/var/lib/kubelet/pods/8781d31c-4701-4e6b-8065-495f1e890c9f/volumes" Feb 27 16:28:09 crc kubenswrapper[4814]: I0227 16:28:09.025284 4814 generic.go:334] "Generic (PLEG): container finished" podID="dd3bb66c-e36b-456f-8051-509099d63405" containerID="81ef0dcfb1aa0d2d1bba5a21ad84b96dea5ed110e08c21309ab4e1e86a512043" exitCode=0 Feb 27 16:28:09 crc kubenswrapper[4814]: I0227 16:28:09.025355 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536828-g5rhp" event={"ID":"dd3bb66c-e36b-456f-8051-509099d63405","Type":"ContainerDied","Data":"81ef0dcfb1aa0d2d1bba5a21ad84b96dea5ed110e08c21309ab4e1e86a512043"} Feb 27 16:28:09 crc kubenswrapper[4814]: I0227 16:28:09.027931 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c5wns" event={"ID":"d3a647cf-eb96-40b8-982c-8e8b869e57c4","Type":"ContainerStarted","Data":"030f38034846ad1b90c3cc6c3063752b224b00566ce8bb90c8792fa36612e952"} Feb 27 16:28:09 crc kubenswrapper[4814]: I0227 16:28:09.073812 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c5wns" podStartSLOduration=2.898089474 podStartE2EDuration="1m10.073790819s" podCreationTimestamp="2026-02-27 16:26:59 +0000 UTC" firstStartedPulling="2026-02-27 16:27:01.220016246 +0000 UTC m=+233.672641076" lastFinishedPulling="2026-02-27 16:28:08.395717581 +0000 UTC m=+300.848342421" observedRunningTime="2026-02-27 16:28:09.072103312 +0000 UTC m=+301.524728142" watchObservedRunningTime="2026-02-27 16:28:09.073790819 +0000 UTC m=+301.526415639" Feb 27 16:28:09 crc kubenswrapper[4814]: I0227 16:28:09.298203 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:28:09 crc kubenswrapper[4814]: I0227 16:28:09.356662 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:28:09 crc kubenswrapper[4814]: I0227 16:28:09.432912 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:28:09 crc kubenswrapper[4814]: I0227 16:28:09.432972 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:28:09 crc kubenswrapper[4814]: I0227 16:28:09.625931 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:28:09 crc kubenswrapper[4814]: I0227 16:28:09.626007 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:28:10 crc kubenswrapper[4814]: I0227 16:28:10.491848 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-sk6rs" podUID="41b851a0-9bda-4e20-a692-5f796f55a749" containerName="registry-server" probeResult="failure" output=< Feb 27 16:28:10 crc kubenswrapper[4814]: timeout: failed to connect service ":50051" within 1s Feb 27 16:28:10 crc kubenswrapper[4814]: > Feb 27 16:28:10 crc kubenswrapper[4814]: I0227 16:28:10.494870 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536828-g5rhp" Feb 27 16:28:10 crc kubenswrapper[4814]: I0227 16:28:10.645788 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cg25x\" (UniqueName: \"kubernetes.io/projected/dd3bb66c-e36b-456f-8051-509099d63405-kube-api-access-cg25x\") pod \"dd3bb66c-e36b-456f-8051-509099d63405\" (UID: \"dd3bb66c-e36b-456f-8051-509099d63405\") " Feb 27 16:28:10 crc kubenswrapper[4814]: I0227 16:28:10.656893 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd3bb66c-e36b-456f-8051-509099d63405-kube-api-access-cg25x" (OuterVolumeSpecName: "kube-api-access-cg25x") pod "dd3bb66c-e36b-456f-8051-509099d63405" (UID: "dd3bb66c-e36b-456f-8051-509099d63405"). InnerVolumeSpecName "kube-api-access-cg25x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:28:10 crc kubenswrapper[4814]: I0227 16:28:10.686053 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-c5wns" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" containerName="registry-server" probeResult="failure" output=< Feb 27 16:28:10 crc kubenswrapper[4814]: timeout: failed to connect service ":50051" within 1s Feb 27 16:28:10 crc kubenswrapper[4814]: > Feb 27 16:28:10 crc kubenswrapper[4814]: I0227 16:28:10.747344 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cg25x\" (UniqueName: \"kubernetes.io/projected/dd3bb66c-e36b-456f-8051-509099d63405-kube-api-access-cg25x\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:11 crc kubenswrapper[4814]: I0227 16:28:11.045010 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536828-g5rhp" event={"ID":"dd3bb66c-e36b-456f-8051-509099d63405","Type":"ContainerDied","Data":"bc91ba7d801bdcb10fd331735b5ffffecb3150f9b3ab387e8a0358818bb4cf99"} Feb 27 16:28:11 crc kubenswrapper[4814]: I0227 16:28:11.045069 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc91ba7d801bdcb10fd331735b5ffffecb3150f9b3ab387e8a0358818bb4cf99" Feb 27 16:28:11 crc kubenswrapper[4814]: I0227 16:28:11.045065 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536828-g5rhp" Feb 27 16:28:12 crc kubenswrapper[4814]: I0227 16:28:12.406835 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:28:12 crc kubenswrapper[4814]: I0227 16:28:12.407621 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:28:13 crc kubenswrapper[4814]: I0227 16:28:13.473392 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9ftgk" podUID="b80cc2cd-d07d-4225-b078-9c70d8e189c7" containerName="registry-server" probeResult="failure" output=< Feb 27 16:28:13 crc kubenswrapper[4814]: timeout: failed to connect service ":50051" within 1s Feb 27 16:28:13 crc kubenswrapper[4814]: > Feb 27 16:28:15 crc kubenswrapper[4814]: I0227 16:28:15.811598 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-565994456b-sh4t6"] Feb 27 16:28:15 crc kubenswrapper[4814]: I0227 16:28:15.812732 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" podUID="1ec92faf-a8c8-490f-95aa-caf7c3e00a2b" containerName="controller-manager" containerID="cri-o://a02afdd51ee070748e9b691122356254428d07f1f13a26a500446c428a0c0f2e" gracePeriod=30 Feb 27 16:28:15 crc kubenswrapper[4814]: I0227 16:28:15.909904 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx"] Feb 27 16:28:15 crc kubenswrapper[4814]: I0227 16:28:15.910591 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" podUID="0801bcb0-748f-4736-ad86-697aff3c5a10" containerName="route-controller-manager" containerID="cri-o://e0128b6e8eb899bee164f8bb852738704d922a765b2f5e4bd8332715f13fecc3" gracePeriod=30 Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.096011 4814 generic.go:334] "Generic (PLEG): container finished" podID="1ec92faf-a8c8-490f-95aa-caf7c3e00a2b" containerID="a02afdd51ee070748e9b691122356254428d07f1f13a26a500446c428a0c0f2e" exitCode=0 Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.096082 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" event={"ID":"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b","Type":"ContainerDied","Data":"a02afdd51ee070748e9b691122356254428d07f1f13a26a500446c428a0c0f2e"} Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.098923 4814 generic.go:334] "Generic (PLEG): container finished" podID="0801bcb0-748f-4736-ad86-697aff3c5a10" containerID="e0128b6e8eb899bee164f8bb852738704d922a765b2f5e4bd8332715f13fecc3" exitCode=0 Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.098987 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" event={"ID":"0801bcb0-748f-4736-ad86-697aff3c5a10","Type":"ContainerDied","Data":"e0128b6e8eb899bee164f8bb852738704d922a765b2f5e4bd8332715f13fecc3"} Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.458667 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.512467 4814 patch_prober.go:28] interesting pod/controller-manager-565994456b-sh4t6 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.62:8443/healthz\": dial tcp 10.217.0.62:8443: connect: connection refused" start-of-body= Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.512574 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" podUID="1ec92faf-a8c8-490f-95aa-caf7c3e00a2b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.62:8443/healthz\": dial tcp 10.217.0.62:8443: connect: connection refused" Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.557804 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr8tb\" (UniqueName: \"kubernetes.io/projected/0801bcb0-748f-4736-ad86-697aff3c5a10-kube-api-access-qr8tb\") pod \"0801bcb0-748f-4736-ad86-697aff3c5a10\" (UID: \"0801bcb0-748f-4736-ad86-697aff3c5a10\") " Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.557847 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0801bcb0-748f-4736-ad86-697aff3c5a10-serving-cert\") pod \"0801bcb0-748f-4736-ad86-697aff3c5a10\" (UID: \"0801bcb0-748f-4736-ad86-697aff3c5a10\") " Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.557882 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0801bcb0-748f-4736-ad86-697aff3c5a10-client-ca\") pod \"0801bcb0-748f-4736-ad86-697aff3c5a10\" (UID: \"0801bcb0-748f-4736-ad86-697aff3c5a10\") " Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.557913 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0801bcb0-748f-4736-ad86-697aff3c5a10-config\") pod \"0801bcb0-748f-4736-ad86-697aff3c5a10\" (UID: \"0801bcb0-748f-4736-ad86-697aff3c5a10\") " Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.559918 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0801bcb0-748f-4736-ad86-697aff3c5a10-client-ca" (OuterVolumeSpecName: "client-ca") pod "0801bcb0-748f-4736-ad86-697aff3c5a10" (UID: "0801bcb0-748f-4736-ad86-697aff3c5a10"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.560834 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0801bcb0-748f-4736-ad86-697aff3c5a10-config" (OuterVolumeSpecName: "config") pod "0801bcb0-748f-4736-ad86-697aff3c5a10" (UID: "0801bcb0-748f-4736-ad86-697aff3c5a10"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.568717 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0801bcb0-748f-4736-ad86-697aff3c5a10-kube-api-access-qr8tb" (OuterVolumeSpecName: "kube-api-access-qr8tb") pod "0801bcb0-748f-4736-ad86-697aff3c5a10" (UID: "0801bcb0-748f-4736-ad86-697aff3c5a10"). InnerVolumeSpecName "kube-api-access-qr8tb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.569663 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0801bcb0-748f-4736-ad86-697aff3c5a10-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0801bcb0-748f-4736-ad86-697aff3c5a10" (UID: "0801bcb0-748f-4736-ad86-697aff3c5a10"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.660390 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr8tb\" (UniqueName: \"kubernetes.io/projected/0801bcb0-748f-4736-ad86-697aff3c5a10-kube-api-access-qr8tb\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.660450 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0801bcb0-748f-4736-ad86-697aff3c5a10-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.660471 4814 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0801bcb0-748f-4736-ad86-697aff3c5a10-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.660488 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0801bcb0-748f-4736-ad86-697aff3c5a10-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.918894 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.965579 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kqn8\" (UniqueName: \"kubernetes.io/projected/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-kube-api-access-7kqn8\") pod \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.965704 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-config\") pod \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.965771 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-client-ca\") pod \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.965899 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-proxy-ca-bundles\") pod \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.965964 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-serving-cert\") pod \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\" (UID: \"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b\") " Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.967091 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1ec92faf-a8c8-490f-95aa-caf7c3e00a2b" (UID: "1ec92faf-a8c8-490f-95aa-caf7c3e00a2b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.967120 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-client-ca" (OuterVolumeSpecName: "client-ca") pod "1ec92faf-a8c8-490f-95aa-caf7c3e00a2b" (UID: "1ec92faf-a8c8-490f-95aa-caf7c3e00a2b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.967536 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-config" (OuterVolumeSpecName: "config") pod "1ec92faf-a8c8-490f-95aa-caf7c3e00a2b" (UID: "1ec92faf-a8c8-490f-95aa-caf7c3e00a2b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.972947 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-kube-api-access-7kqn8" (OuterVolumeSpecName: "kube-api-access-7kqn8") pod "1ec92faf-a8c8-490f-95aa-caf7c3e00a2b" (UID: "1ec92faf-a8c8-490f-95aa-caf7c3e00a2b"). InnerVolumeSpecName "kube-api-access-7kqn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:28:16 crc kubenswrapper[4814]: I0227 16:28:16.973290 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1ec92faf-a8c8-490f-95aa-caf7c3e00a2b" (UID: "1ec92faf-a8c8-490f-95aa-caf7c3e00a2b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.068810 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.068872 4814 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.068893 4814 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.068916 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.068945 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kqn8\" (UniqueName: \"kubernetes.io/projected/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b-kube-api-access-7kqn8\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.114856 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" event={"ID":"0801bcb0-748f-4736-ad86-697aff3c5a10","Type":"ContainerDied","Data":"4790a09a578f52cc4f4c55b1a49b04f58fa2d8ec15705564c09477137d256e8d"} Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.114946 4814 scope.go:117] "RemoveContainer" containerID="e0128b6e8eb899bee164f8bb852738704d922a765b2f5e4bd8332715f13fecc3" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.115122 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.120327 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" event={"ID":"1ec92faf-a8c8-490f-95aa-caf7c3e00a2b","Type":"ContainerDied","Data":"6259c3bb13bc35ad0d8d8c9dde16c8d243950febe10face485593554186b827e"} Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.120468 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-565994456b-sh4t6" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.144059 4814 scope.go:117] "RemoveContainer" containerID="a02afdd51ee070748e9b691122356254428d07f1f13a26a500446c428a0c0f2e" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.175792 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx"] Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.180667 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54f96c65c5-chjlx"] Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.187180 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-565994456b-sh4t6"] Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.201785 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-565994456b-sh4t6"] Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.543501 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-689bc584d8-f9q99"] Feb 27 16:28:17 crc kubenswrapper[4814]: E0227 16:28:17.543952 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0801bcb0-748f-4736-ad86-697aff3c5a10" containerName="route-controller-manager" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.543978 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="0801bcb0-748f-4736-ad86-697aff3c5a10" containerName="route-controller-manager" Feb 27 16:28:17 crc kubenswrapper[4814]: E0227 16:28:17.544012 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8781d31c-4701-4e6b-8065-495f1e890c9f" containerName="extract-utilities" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.544026 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="8781d31c-4701-4e6b-8065-495f1e890c9f" containerName="extract-utilities" Feb 27 16:28:17 crc kubenswrapper[4814]: E0227 16:28:17.544049 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cdfdb15-014d-40fc-9a7b-a905278f1249" containerName="extract-utilities" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.544064 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cdfdb15-014d-40fc-9a7b-a905278f1249" containerName="extract-utilities" Feb 27 16:28:17 crc kubenswrapper[4814]: E0227 16:28:17.544086 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cdfdb15-014d-40fc-9a7b-a905278f1249" containerName="extract-content" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.544099 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cdfdb15-014d-40fc-9a7b-a905278f1249" containerName="extract-content" Feb 27 16:28:17 crc kubenswrapper[4814]: E0227 16:28:17.544123 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cdfdb15-014d-40fc-9a7b-a905278f1249" containerName="registry-server" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.544138 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cdfdb15-014d-40fc-9a7b-a905278f1249" containerName="registry-server" Feb 27 16:28:17 crc kubenswrapper[4814]: E0227 16:28:17.544158 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd3bb66c-e36b-456f-8051-509099d63405" containerName="oc" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.544171 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd3bb66c-e36b-456f-8051-509099d63405" containerName="oc" Feb 27 16:28:17 crc kubenswrapper[4814]: E0227 16:28:17.544188 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8781d31c-4701-4e6b-8065-495f1e890c9f" containerName="extract-content" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.544200 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="8781d31c-4701-4e6b-8065-495f1e890c9f" containerName="extract-content" Feb 27 16:28:17 crc kubenswrapper[4814]: E0227 16:28:17.544214 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ec92faf-a8c8-490f-95aa-caf7c3e00a2b" containerName="controller-manager" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.544228 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ec92faf-a8c8-490f-95aa-caf7c3e00a2b" containerName="controller-manager" Feb 27 16:28:17 crc kubenswrapper[4814]: E0227 16:28:17.544250 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8781d31c-4701-4e6b-8065-495f1e890c9f" containerName="registry-server" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.544294 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="8781d31c-4701-4e6b-8065-495f1e890c9f" containerName="registry-server" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.544482 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cdfdb15-014d-40fc-9a7b-a905278f1249" containerName="registry-server" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.544508 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd3bb66c-e36b-456f-8051-509099d63405" containerName="oc" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.544535 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="8781d31c-4701-4e6b-8065-495f1e890c9f" containerName="registry-server" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.544551 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="0801bcb0-748f-4736-ad86-697aff3c5a10" containerName="route-controller-manager" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.544570 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ec92faf-a8c8-490f-95aa-caf7c3e00a2b" containerName="controller-manager" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.545351 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.554583 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.557195 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.557314 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.557833 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz"] Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.558415 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.559367 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.558507 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.560719 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.572716 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.573083 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.573644 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.573702 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.573880 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.574185 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.575363 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-proxy-ca-bundles\") pod \"controller-manager-689bc584d8-f9q99\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.575476 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-config\") pod \"controller-manager-689bc584d8-f9q99\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.575569 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-config\") pod \"route-controller-manager-69bc6f769b-bfkvz\" (UID: \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\") " pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.575679 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-client-ca\") pod \"controller-manager-689bc584d8-f9q99\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.575750 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3297f6be-fd91-4f82-8d3f-6e1fe627b449-serving-cert\") pod \"controller-manager-689bc584d8-f9q99\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.575830 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-client-ca\") pod \"route-controller-manager-69bc6f769b-bfkvz\" (UID: \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\") " pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.575924 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f7gh\" (UniqueName: \"kubernetes.io/projected/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-kube-api-access-5f7gh\") pod \"route-controller-manager-69bc6f769b-bfkvz\" (UID: \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\") " pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.576017 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b278\" (UniqueName: \"kubernetes.io/projected/3297f6be-fd91-4f82-8d3f-6e1fe627b449-kube-api-access-5b278\") pod \"controller-manager-689bc584d8-f9q99\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.576107 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-serving-cert\") pod \"route-controller-manager-69bc6f769b-bfkvz\" (UID: \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\") " pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.576547 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.580511 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-689bc584d8-f9q99"] Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.589318 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz"] Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.678549 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b278\" (UniqueName: \"kubernetes.io/projected/3297f6be-fd91-4f82-8d3f-6e1fe627b449-kube-api-access-5b278\") pod \"controller-manager-689bc584d8-f9q99\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.679089 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-serving-cert\") pod \"route-controller-manager-69bc6f769b-bfkvz\" (UID: \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\") " pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.679299 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-proxy-ca-bundles\") pod \"controller-manager-689bc584d8-f9q99\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.679429 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-config\") pod \"controller-manager-689bc584d8-f9q99\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.679562 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-config\") pod \"route-controller-manager-69bc6f769b-bfkvz\" (UID: \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\") " pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.679778 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-client-ca\") pod \"controller-manager-689bc584d8-f9q99\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.679910 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3297f6be-fd91-4f82-8d3f-6e1fe627b449-serving-cert\") pod \"controller-manager-689bc584d8-f9q99\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.680031 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-client-ca\") pod \"route-controller-manager-69bc6f769b-bfkvz\" (UID: \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\") " pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.680163 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f7gh\" (UniqueName: \"kubernetes.io/projected/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-kube-api-access-5f7gh\") pod \"route-controller-manager-69bc6f769b-bfkvz\" (UID: \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\") " pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.681060 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-config\") pod \"controller-manager-689bc584d8-f9q99\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.686626 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-proxy-ca-bundles\") pod \"controller-manager-689bc584d8-f9q99\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.687230 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-config\") pod \"route-controller-manager-69bc6f769b-bfkvz\" (UID: \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\") " pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.690945 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-serving-cert\") pod \"route-controller-manager-69bc6f769b-bfkvz\" (UID: \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\") " pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.695863 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-client-ca\") pod \"controller-manager-689bc584d8-f9q99\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.696994 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-client-ca\") pod \"route-controller-manager-69bc6f769b-bfkvz\" (UID: \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\") " pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.700369 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3297f6be-fd91-4f82-8d3f-6e1fe627b449-serving-cert\") pod \"controller-manager-689bc584d8-f9q99\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.705482 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b278\" (UniqueName: \"kubernetes.io/projected/3297f6be-fd91-4f82-8d3f-6e1fe627b449-kube-api-access-5b278\") pod \"controller-manager-689bc584d8-f9q99\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.705672 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f7gh\" (UniqueName: \"kubernetes.io/projected/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-kube-api-access-5f7gh\") pod \"route-controller-manager-69bc6f769b-bfkvz\" (UID: \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\") " pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.873650 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:17 crc kubenswrapper[4814]: I0227 16:28:17.892193 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:28:18 crc kubenswrapper[4814]: I0227 16:28:18.105497 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-689bc584d8-f9q99"] Feb 27 16:28:18 crc kubenswrapper[4814]: I0227 16:28:18.132054 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" event={"ID":"3297f6be-fd91-4f82-8d3f-6e1fe627b449","Type":"ContainerStarted","Data":"ead29187da7fda139458e1ba01394e20dc3d27538231ed1227798c87069d0751"} Feb 27 16:28:18 crc kubenswrapper[4814]: I0227 16:28:18.173104 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz"] Feb 27 16:28:18 crc kubenswrapper[4814]: I0227 16:28:18.496355 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0801bcb0-748f-4736-ad86-697aff3c5a10" path="/var/lib/kubelet/pods/0801bcb0-748f-4736-ad86-697aff3c5a10/volumes" Feb 27 16:28:18 crc kubenswrapper[4814]: I0227 16:28:18.497465 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ec92faf-a8c8-490f-95aa-caf7c3e00a2b" path="/var/lib/kubelet/pods/1ec92faf-a8c8-490f-95aa-caf7c3e00a2b/volumes" Feb 27 16:28:19 crc kubenswrapper[4814]: I0227 16:28:19.279411 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" event={"ID":"10fa0a49-7f2e-4001-a99a-29ec8ce7c921","Type":"ContainerStarted","Data":"d70877336baf2a47b41a630c1f36e75d576381e67a5869647a77e9d1f17e9a43"} Feb 27 16:28:19 crc kubenswrapper[4814]: I0227 16:28:19.279457 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" event={"ID":"10fa0a49-7f2e-4001-a99a-29ec8ce7c921","Type":"ContainerStarted","Data":"ceb5c785cf42ef9597bf9140bb56254a806c04027dad1c0033f079269a54f481"} Feb 27 16:28:19 crc kubenswrapper[4814]: I0227 16:28:19.279728 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:28:19 crc kubenswrapper[4814]: I0227 16:28:19.288719 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" event={"ID":"3297f6be-fd91-4f82-8d3f-6e1fe627b449","Type":"ContainerStarted","Data":"d52b0abb8f8cbbfa2859761e99c665f2fed9ad8b6ff546737caaee892c4d4731"} Feb 27 16:28:19 crc kubenswrapper[4814]: I0227 16:28:19.289559 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:19 crc kubenswrapper[4814]: I0227 16:28:19.305114 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:28:19 crc kubenswrapper[4814]: I0227 16:28:19.308874 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" podStartSLOduration=4.3088487109999996 podStartE2EDuration="4.308848711s" podCreationTimestamp="2026-02-27 16:28:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:28:19.305248808 +0000 UTC m=+311.757873638" watchObservedRunningTime="2026-02-27 16:28:19.308848711 +0000 UTC m=+311.761473541" Feb 27 16:28:19 crc kubenswrapper[4814]: I0227 16:28:19.320815 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:28:19 crc kubenswrapper[4814]: I0227 16:28:19.324653 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" podStartSLOduration=4.32463126 podStartE2EDuration="4.32463126s" podCreationTimestamp="2026-02-27 16:28:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:28:19.321369449 +0000 UTC m=+311.773994299" watchObservedRunningTime="2026-02-27 16:28:19.32463126 +0000 UTC m=+311.777256090" Feb 27 16:28:19 crc kubenswrapper[4814]: I0227 16:28:19.475517 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:28:19 crc kubenswrapper[4814]: I0227 16:28:19.514299 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:28:19 crc kubenswrapper[4814]: I0227 16:28:19.665403 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:28:19 crc kubenswrapper[4814]: I0227 16:28:19.713560 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:28:21 crc kubenswrapper[4814]: I0227 16:28:21.104105 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-k4p69"] Feb 27 16:28:21 crc kubenswrapper[4814]: I0227 16:28:21.849594 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c5wns"] Feb 27 16:28:21 crc kubenswrapper[4814]: I0227 16:28:21.849972 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c5wns" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" containerName="registry-server" containerID="cri-o://030f38034846ad1b90c3cc6c3063752b224b00566ce8bb90c8792fa36612e952" gracePeriod=2 Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.311989 4814 generic.go:334] "Generic (PLEG): container finished" podID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" containerID="030f38034846ad1b90c3cc6c3063752b224b00566ce8bb90c8792fa36612e952" exitCode=0 Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.312467 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c5wns" event={"ID":"d3a647cf-eb96-40b8-982c-8e8b869e57c4","Type":"ContainerDied","Data":"030f38034846ad1b90c3cc6c3063752b224b00566ce8bb90c8792fa36612e952"} Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.360985 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.451439 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.507008 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.524762 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a647cf-eb96-40b8-982c-8e8b869e57c4-catalog-content\") pod \"d3a647cf-eb96-40b8-982c-8e8b869e57c4\" (UID: \"d3a647cf-eb96-40b8-982c-8e8b869e57c4\") " Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.524836 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a647cf-eb96-40b8-982c-8e8b869e57c4-utilities\") pod \"d3a647cf-eb96-40b8-982c-8e8b869e57c4\" (UID: \"d3a647cf-eb96-40b8-982c-8e8b869e57c4\") " Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.524868 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rddgl\" (UniqueName: \"kubernetes.io/projected/d3a647cf-eb96-40b8-982c-8e8b869e57c4-kube-api-access-rddgl\") pod \"d3a647cf-eb96-40b8-982c-8e8b869e57c4\" (UID: \"d3a647cf-eb96-40b8-982c-8e8b869e57c4\") " Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.526068 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3a647cf-eb96-40b8-982c-8e8b869e57c4-utilities" (OuterVolumeSpecName: "utilities") pod "d3a647cf-eb96-40b8-982c-8e8b869e57c4" (UID: "d3a647cf-eb96-40b8-982c-8e8b869e57c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.533982 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3a647cf-eb96-40b8-982c-8e8b869e57c4-kube-api-access-rddgl" (OuterVolumeSpecName: "kube-api-access-rddgl") pod "d3a647cf-eb96-40b8-982c-8e8b869e57c4" (UID: "d3a647cf-eb96-40b8-982c-8e8b869e57c4"). InnerVolumeSpecName "kube-api-access-rddgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.584017 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3a647cf-eb96-40b8-982c-8e8b869e57c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3a647cf-eb96-40b8-982c-8e8b869e57c4" (UID: "d3a647cf-eb96-40b8-982c-8e8b869e57c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.626641 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a647cf-eb96-40b8-982c-8e8b869e57c4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.626680 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a647cf-eb96-40b8-982c-8e8b869e57c4-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.626690 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rddgl\" (UniqueName: \"kubernetes.io/projected/d3a647cf-eb96-40b8-982c-8e8b869e57c4-kube-api-access-rddgl\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.902510 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.902945 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.903007 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.903999 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223"} pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.904083 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" containerID="cri-o://c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223" gracePeriod=600 Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.935874 4814 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.936283 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479" gracePeriod=15 Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.936386 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced" gracePeriod=15 Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.936413 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680" gracePeriod=15 Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.936411 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5" gracePeriod=15 Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.936408 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51" gracePeriod=15 Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.939947 4814 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 27 16:28:22 crc kubenswrapper[4814]: E0227 16:28:22.940220 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940239 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 27 16:28:22 crc kubenswrapper[4814]: E0227 16:28:22.940265 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940274 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 16:28:22 crc kubenswrapper[4814]: E0227 16:28:22.940284 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940291 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 16:28:22 crc kubenswrapper[4814]: E0227 16:28:22.940300 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" containerName="extract-content" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940305 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" containerName="extract-content" Feb 27 16:28:22 crc kubenswrapper[4814]: E0227 16:28:22.940317 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" containerName="registry-server" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940335 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" containerName="registry-server" Feb 27 16:28:22 crc kubenswrapper[4814]: E0227 16:28:22.940343 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940350 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 27 16:28:22 crc kubenswrapper[4814]: E0227 16:28:22.940357 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940364 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 27 16:28:22 crc kubenswrapper[4814]: E0227 16:28:22.940377 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" containerName="extract-utilities" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940384 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" containerName="extract-utilities" Feb 27 16:28:22 crc kubenswrapper[4814]: E0227 16:28:22.940397 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940403 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 16:28:22 crc kubenswrapper[4814]: E0227 16:28:22.940413 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940422 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 27 16:28:22 crc kubenswrapper[4814]: E0227 16:28:22.940430 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940438 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 16:28:22 crc kubenswrapper[4814]: E0227 16:28:22.940447 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940455 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 16:28:22 crc kubenswrapper[4814]: E0227 16:28:22.940464 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940471 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940595 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940605 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" containerName="registry-server" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940616 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940628 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940637 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940647 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940656 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940664 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.940892 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.941104 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.942010 4814 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.942504 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.948624 4814 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Feb 27 16:28:22 crc kubenswrapper[4814]: I0227 16:28:22.991197 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.033193 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.033308 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.033338 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.033392 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.033421 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.033468 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.033501 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.033525 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.134718 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.134794 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.134820 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.134868 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.134876 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.134943 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.134893 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.134967 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.135000 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.135041 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.135138 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.135215 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.135231 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.135279 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.135325 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.135286 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.289446 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.324618 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.326321 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.327071 4814 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680" exitCode=0 Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.327103 4814 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51" exitCode=0 Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.327114 4814 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced" exitCode=0 Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.327123 4814 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5" exitCode=2 Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.327203 4814 scope.go:117] "RemoveContainer" containerID="bce056f1325a3b3dc3ce691eaad73f2caa35081e5bc200a171ea8596e12883c1" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.330220 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c5wns" event={"ID":"d3a647cf-eb96-40b8-982c-8e8b869e57c4","Type":"ContainerDied","Data":"c18a0d2234e658205a945f939203e622c10efd7c68e81f88e2e5f8143bebbc7e"} Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.330371 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c5wns" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.334539 4814 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.338196 4814 status_manager.go:851] "Failed to get status for pod" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" pod="openshift-marketplace/certified-operators-c5wns" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c5wns\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.339288 4814 generic.go:334] "Generic (PLEG): container finished" podID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerID="c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223" exitCode=0 Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.339883 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerDied","Data":"c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223"} Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.339913 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"7d5c0ac92b2e6bacf90ee7b6e5821df1c3f5a2b33024a13ffeb8403e774e5521"} Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.340770 4814 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.341105 4814 status_manager.go:851] "Failed to get status for pod" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" pod="openshift-marketplace/certified-operators-c5wns" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c5wns\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.341525 4814 status_manager.go:851] "Failed to get status for pod" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-nlbk2\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.351086 4814 status_manager.go:851] "Failed to get status for pod" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" pod="openshift-marketplace/certified-operators-c5wns" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c5wns\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.351749 4814 status_manager.go:851] "Failed to get status for pod" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-nlbk2\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.352504 4814 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.373358 4814 scope.go:117] "RemoveContainer" containerID="030f38034846ad1b90c3cc6c3063752b224b00566ce8bb90c8792fa36612e952" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.397332 4814 scope.go:117] "RemoveContainer" containerID="1899b6570d51daf80aa21a3f009ab53315427b24f79c18be45988a8e74237c7e" Feb 27 16:28:23 crc kubenswrapper[4814]: I0227 16:28:23.419369 4814 scope.go:117] "RemoveContainer" containerID="3749fb0334366dc48359430c6354be4ed08fcf1840c7a770d9874cfb029e4a0f" Feb 27 16:28:24 crc kubenswrapper[4814]: I0227 16:28:24.353964 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"57009b3e63779816d725ceea035b1b0f332445201671efa15bacdd10da79e48a"} Feb 27 16:28:24 crc kubenswrapper[4814]: I0227 16:28:24.354377 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"0899796d20352dc2719d54226db1127842a8e039b9f672e1b3091deac6e88c75"} Feb 27 16:28:24 crc kubenswrapper[4814]: I0227 16:28:24.356787 4814 status_manager.go:851] "Failed to get status for pod" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" pod="openshift-marketplace/certified-operators-c5wns" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c5wns\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:24 crc kubenswrapper[4814]: I0227 16:28:24.357863 4814 status_manager.go:851] "Failed to get status for pod" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-nlbk2\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:24 crc kubenswrapper[4814]: I0227 16:28:24.359125 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 16:28:24 crc kubenswrapper[4814]: I0227 16:28:24.359350 4814 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:25 crc kubenswrapper[4814]: I0227 16:28:25.510284 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 16:28:25 crc kubenswrapper[4814]: I0227 16:28:25.511983 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:25 crc kubenswrapper[4814]: I0227 16:28:25.513392 4814 status_manager.go:851] "Failed to get status for pod" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" pod="openshift-marketplace/certified-operators-c5wns" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c5wns\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:25 crc kubenswrapper[4814]: I0227 16:28:25.513736 4814 status_manager.go:851] "Failed to get status for pod" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-nlbk2\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:25 crc kubenswrapper[4814]: I0227 16:28:25.514354 4814 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:25 crc kubenswrapper[4814]: I0227 16:28:25.515024 4814 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:25 crc kubenswrapper[4814]: I0227 16:28:25.679386 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 27 16:28:25 crc kubenswrapper[4814]: I0227 16:28:25.679498 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 27 16:28:25 crc kubenswrapper[4814]: I0227 16:28:25.679548 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 27 16:28:25 crc kubenswrapper[4814]: I0227 16:28:25.679542 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:28:25 crc kubenswrapper[4814]: I0227 16:28:25.679653 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:28:25 crc kubenswrapper[4814]: I0227 16:28:25.679681 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:28:25 crc kubenswrapper[4814]: I0227 16:28:25.680195 4814 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:25 crc kubenswrapper[4814]: I0227 16:28:25.680234 4814 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:25 crc kubenswrapper[4814]: I0227 16:28:25.680307 4814 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.389673 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.392007 4814 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479" exitCode=0 Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.392090 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.392223 4814 scope.go:117] "RemoveContainer" containerID="db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.411738 4814 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.412305 4814 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.412751 4814 status_manager.go:851] "Failed to get status for pod" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" pod="openshift-marketplace/certified-operators-c5wns" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c5wns\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.413191 4814 status_manager.go:851] "Failed to get status for pod" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-nlbk2\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.419812 4814 scope.go:117] "RemoveContainer" containerID="2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.438898 4814 scope.go:117] "RemoveContainer" containerID="da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.461233 4814 scope.go:117] "RemoveContainer" containerID="cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.482249 4814 scope.go:117] "RemoveContainer" containerID="76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.501337 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.506792 4814 scope.go:117] "RemoveContainer" containerID="6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.530036 4814 scope.go:117] "RemoveContainer" containerID="db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680" Feb 27 16:28:26 crc kubenswrapper[4814]: E0227 16:28:26.531713 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\": container with ID starting with db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680 not found: ID does not exist" containerID="db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.531754 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680"} err="failed to get container status \"db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\": rpc error: code = NotFound desc = could not find container \"db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680\": container with ID starting with db7437b84b1ad7293ef1c1486fbda52de8e53b4e1a7abe8d1b10e0e3a073a680 not found: ID does not exist" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.531788 4814 scope.go:117] "RemoveContainer" containerID="2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51" Feb 27 16:28:26 crc kubenswrapper[4814]: E0227 16:28:26.531988 4814 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.189:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" volumeName="registry-storage" Feb 27 16:28:26 crc kubenswrapper[4814]: E0227 16:28:26.532227 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\": container with ID starting with 2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51 not found: ID does not exist" containerID="2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.532332 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51"} err="failed to get container status \"2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\": rpc error: code = NotFound desc = could not find container \"2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51\": container with ID starting with 2d6619dfe916bf5beee8cf9f39fe5394b1ca30e79b872408f076a668a7d2cf51 not found: ID does not exist" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.532350 4814 scope.go:117] "RemoveContainer" containerID="da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced" Feb 27 16:28:26 crc kubenswrapper[4814]: E0227 16:28:26.532677 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\": container with ID starting with da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced not found: ID does not exist" containerID="da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.532728 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced"} err="failed to get container status \"da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\": rpc error: code = NotFound desc = could not find container \"da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced\": container with ID starting with da70fe59c0aee8b733d57d5723e71e7f9bbaa55c4a7ccaa27ef9cf0a62956ced not found: ID does not exist" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.532764 4814 scope.go:117] "RemoveContainer" containerID="cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5" Feb 27 16:28:26 crc kubenswrapper[4814]: E0227 16:28:26.533120 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\": container with ID starting with cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5 not found: ID does not exist" containerID="cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.533140 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5"} err="failed to get container status \"cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\": rpc error: code = NotFound desc = could not find container \"cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5\": container with ID starting with cbf73bd859d04e385d5ab143536690c10f1b5553afc7b9c55e9fbf125d68cec5 not found: ID does not exist" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.533155 4814 scope.go:117] "RemoveContainer" containerID="76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479" Feb 27 16:28:26 crc kubenswrapper[4814]: E0227 16:28:26.533517 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\": container with ID starting with 76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479 not found: ID does not exist" containerID="76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.533535 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479"} err="failed to get container status \"76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\": rpc error: code = NotFound desc = could not find container \"76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479\": container with ID starting with 76f67fdd4617aee5a02eb2c6656147a50792db0268cb27234ccbbf1f25193479 not found: ID does not exist" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.533549 4814 scope.go:117] "RemoveContainer" containerID="6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82" Feb 27 16:28:26 crc kubenswrapper[4814]: E0227 16:28:26.533871 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\": container with ID starting with 6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82 not found: ID does not exist" containerID="6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82" Feb 27 16:28:26 crc kubenswrapper[4814]: I0227 16:28:26.533901 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82"} err="failed to get container status \"6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\": rpc error: code = NotFound desc = could not find container \"6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82\": container with ID starting with 6d7794090bf6a26d40297522d2efa5cfd6ea804862d7ed5db64cb72b08719c82 not found: ID does not exist" Feb 27 16:28:28 crc kubenswrapper[4814]: E0227 16:28:28.044752 4814 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/events/machine-config-daemon-nlbk2.18982724ed529c43\": dial tcp 38.102.83.189:6443: connect: connection refused" event="&Event{ObjectMeta:{machine-config-daemon-nlbk2.18982724ed529c43 openshift-machine-config-operator 26812 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:machine-config-daemon-nlbk2,UID:be713cf4-11a1-42e7-93c6-ab3f748570a4,APIVersion:v1,ResourceVersion:26805,FieldPath:spec.containers{machine-config-daemon},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:24:52 +0000 UTC,LastTimestamp:2026-02-27 16:28:23.034165099 +0000 UTC m=+315.486789929,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:28:28 crc kubenswrapper[4814]: I0227 16:28:28.412112 4814 generic.go:334] "Generic (PLEG): container finished" podID="306421d3-af6e-43c8-9411-a8eaaf583862" containerID="84352145e637126e0a3480c234b3eae3bd5f7bc9152740430bf73092114a1a90" exitCode=0 Feb 27 16:28:28 crc kubenswrapper[4814]: I0227 16:28:28.412178 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"306421d3-af6e-43c8-9411-a8eaaf583862","Type":"ContainerDied","Data":"84352145e637126e0a3480c234b3eae3bd5f7bc9152740430bf73092114a1a90"} Feb 27 16:28:28 crc kubenswrapper[4814]: I0227 16:28:28.412989 4814 status_manager.go:851] "Failed to get status for pod" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" pod="openshift-marketplace/certified-operators-c5wns" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c5wns\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:28 crc kubenswrapper[4814]: I0227 16:28:28.413400 4814 status_manager.go:851] "Failed to get status for pod" podUID="306421d3-af6e-43c8-9411-a8eaaf583862" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:28 crc kubenswrapper[4814]: I0227 16:28:28.413991 4814 status_manager.go:851] "Failed to get status for pod" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-nlbk2\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:28 crc kubenswrapper[4814]: I0227 16:28:28.414588 4814 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:28 crc kubenswrapper[4814]: I0227 16:28:28.498242 4814 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:28 crc kubenswrapper[4814]: I0227 16:28:28.499964 4814 status_manager.go:851] "Failed to get status for pod" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" pod="openshift-marketplace/certified-operators-c5wns" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c5wns\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:28 crc kubenswrapper[4814]: I0227 16:28:28.500455 4814 status_manager.go:851] "Failed to get status for pod" podUID="306421d3-af6e-43c8-9411-a8eaaf583862" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:28 crc kubenswrapper[4814]: I0227 16:28:28.500765 4814 status_manager.go:851] "Failed to get status for pod" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-nlbk2\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:29 crc kubenswrapper[4814]: I0227 16:28:29.844721 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 27 16:28:29 crc kubenswrapper[4814]: I0227 16:28:29.846605 4814 status_manager.go:851] "Failed to get status for pod" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" pod="openshift-marketplace/certified-operators-c5wns" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c5wns\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:29 crc kubenswrapper[4814]: I0227 16:28:29.847224 4814 status_manager.go:851] "Failed to get status for pod" podUID="306421d3-af6e-43c8-9411-a8eaaf583862" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:29 crc kubenswrapper[4814]: I0227 16:28:29.847829 4814 status_manager.go:851] "Failed to get status for pod" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-nlbk2\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:29 crc kubenswrapper[4814]: I0227 16:28:29.848217 4814 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:29 crc kubenswrapper[4814]: I0227 16:28:29.951579 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/306421d3-af6e-43c8-9411-a8eaaf583862-kubelet-dir\") pod \"306421d3-af6e-43c8-9411-a8eaaf583862\" (UID: \"306421d3-af6e-43c8-9411-a8eaaf583862\") " Feb 27 16:28:29 crc kubenswrapper[4814]: I0227 16:28:29.951723 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/306421d3-af6e-43c8-9411-a8eaaf583862-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "306421d3-af6e-43c8-9411-a8eaaf583862" (UID: "306421d3-af6e-43c8-9411-a8eaaf583862"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:28:29 crc kubenswrapper[4814]: I0227 16:28:29.951760 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/306421d3-af6e-43c8-9411-a8eaaf583862-var-lock\") pod \"306421d3-af6e-43c8-9411-a8eaaf583862\" (UID: \"306421d3-af6e-43c8-9411-a8eaaf583862\") " Feb 27 16:28:29 crc kubenswrapper[4814]: I0227 16:28:29.951861 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/306421d3-af6e-43c8-9411-a8eaaf583862-var-lock" (OuterVolumeSpecName: "var-lock") pod "306421d3-af6e-43c8-9411-a8eaaf583862" (UID: "306421d3-af6e-43c8-9411-a8eaaf583862"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:28:29 crc kubenswrapper[4814]: I0227 16:28:29.951915 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/306421d3-af6e-43c8-9411-a8eaaf583862-kube-api-access\") pod \"306421d3-af6e-43c8-9411-a8eaaf583862\" (UID: \"306421d3-af6e-43c8-9411-a8eaaf583862\") " Feb 27 16:28:29 crc kubenswrapper[4814]: I0227 16:28:29.952386 4814 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/306421d3-af6e-43c8-9411-a8eaaf583862-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:29 crc kubenswrapper[4814]: I0227 16:28:29.952418 4814 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/306421d3-af6e-43c8-9411-a8eaaf583862-var-lock\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:29 crc kubenswrapper[4814]: I0227 16:28:29.960916 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/306421d3-af6e-43c8-9411-a8eaaf583862-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "306421d3-af6e-43c8-9411-a8eaaf583862" (UID: "306421d3-af6e-43c8-9411-a8eaaf583862"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:28:30 crc kubenswrapper[4814]: I0227 16:28:30.054544 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/306421d3-af6e-43c8-9411-a8eaaf583862-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:30 crc kubenswrapper[4814]: E0227 16:28:30.238369 4814 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:30 crc kubenswrapper[4814]: E0227 16:28:30.238944 4814 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:30 crc kubenswrapper[4814]: E0227 16:28:30.239518 4814 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:30 crc kubenswrapper[4814]: E0227 16:28:30.239968 4814 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:30 crc kubenswrapper[4814]: E0227 16:28:30.240486 4814 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:30 crc kubenswrapper[4814]: I0227 16:28:30.240550 4814 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 27 16:28:30 crc kubenswrapper[4814]: E0227 16:28:30.240911 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" interval="200ms" Feb 27 16:28:30 crc kubenswrapper[4814]: I0227 16:28:30.430816 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"306421d3-af6e-43c8-9411-a8eaaf583862","Type":"ContainerDied","Data":"159b163211cc0919ef75fa57e0135ef1a7e25179bbb9a1b69ef98058200a8e7e"} Feb 27 16:28:30 crc kubenswrapper[4814]: I0227 16:28:30.430882 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="159b163211cc0919ef75fa57e0135ef1a7e25179bbb9a1b69ef98058200a8e7e" Feb 27 16:28:30 crc kubenswrapper[4814]: I0227 16:28:30.430913 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 27 16:28:30 crc kubenswrapper[4814]: E0227 16:28:30.441934 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" interval="400ms" Feb 27 16:28:30 crc kubenswrapper[4814]: I0227 16:28:30.455293 4814 status_manager.go:851] "Failed to get status for pod" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" pod="openshift-marketplace/certified-operators-c5wns" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c5wns\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:30 crc kubenswrapper[4814]: I0227 16:28:30.455633 4814 status_manager.go:851] "Failed to get status for pod" podUID="306421d3-af6e-43c8-9411-a8eaaf583862" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:30 crc kubenswrapper[4814]: I0227 16:28:30.456038 4814 status_manager.go:851] "Failed to get status for pod" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-nlbk2\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:30 crc kubenswrapper[4814]: I0227 16:28:30.456561 4814 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:30 crc kubenswrapper[4814]: E0227 16:28:30.843635 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" interval="800ms" Feb 27 16:28:31 crc kubenswrapper[4814]: E0227 16:28:31.645720 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" interval="1.6s" Feb 27 16:28:33 crc kubenswrapper[4814]: E0227 16:28:33.247178 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" interval="3.2s" Feb 27 16:28:36 crc kubenswrapper[4814]: E0227 16:28:36.448942 4814 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" interval="6.4s" Feb 27 16:28:37 crc kubenswrapper[4814]: E0227 16:28:37.301618 4814 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/events/machine-config-daemon-nlbk2.18982724ed529c43\": dial tcp 38.102.83.189:6443: connect: connection refused" event="&Event{ObjectMeta:{machine-config-daemon-nlbk2.18982724ed529c43 openshift-machine-config-operator 26812 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:machine-config-daemon-nlbk2,UID:be713cf4-11a1-42e7-93c6-ab3f748570a4,APIVersion:v1,ResourceVersion:26805,FieldPath:spec.containers{machine-config-daemon},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 16:24:52 +0000 UTC,LastTimestamp:2026-02-27 16:28:23.034165099 +0000 UTC m=+315.486789929,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.486795 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.488937 4814 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.489477 4814 status_manager.go:851] "Failed to get status for pod" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" pod="openshift-marketplace/certified-operators-c5wns" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c5wns\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.490034 4814 status_manager.go:851] "Failed to get status for pod" podUID="306421d3-af6e-43c8-9411-a8eaaf583862" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.490959 4814 status_manager.go:851] "Failed to get status for pod" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-nlbk2\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.496312 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.497192 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.497281 4814 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="501ea405666a49080dd2af178db81a992757c23da06fedd1a6036047880f3f7a" exitCode=1 Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.497317 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"501ea405666a49080dd2af178db81a992757c23da06fedd1a6036047880f3f7a"} Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.497852 4814 scope.go:117] "RemoveContainer" containerID="501ea405666a49080dd2af178db81a992757c23da06fedd1a6036047880f3f7a" Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.498477 4814 status_manager.go:851] "Failed to get status for pod" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-nlbk2\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.499169 4814 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.499599 4814 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.499938 4814 status_manager.go:851] "Failed to get status for pod" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" pod="openshift-marketplace/certified-operators-c5wns" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c5wns\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.500297 4814 status_manager.go:851] "Failed to get status for pod" podUID="306421d3-af6e-43c8-9411-a8eaaf583862" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.514890 4814 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="929bfc6f-c721-4f45-9674-fc3f151c3f6e" Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.514944 4814 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="929bfc6f-c721-4f45-9674-fc3f151c3f6e" Feb 27 16:28:37 crc kubenswrapper[4814]: E0227 16:28:37.515798 4814 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.516718 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:37 crc kubenswrapper[4814]: W0227 16:28:37.544740 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-5426bac6f438de152af29a498c06c72fdd24ef998c5f44f8591b4004b11e7056 WatchSource:0}: Error finding container 5426bac6f438de152af29a498c06c72fdd24ef998c5f44f8591b4004b11e7056: Status 404 returned error can't find the container with id 5426bac6f438de152af29a498c06c72fdd24ef998c5f44f8591b4004b11e7056 Feb 27 16:28:37 crc kubenswrapper[4814]: I0227 16:28:37.688981 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.500057 4814 status_manager.go:851] "Failed to get status for pod" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-nlbk2\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.501031 4814 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.501411 4814 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.502063 4814 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.502916 4814 status_manager.go:851] "Failed to get status for pod" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" pod="openshift-marketplace/certified-operators-c5wns" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c5wns\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.503387 4814 status_manager.go:851] "Failed to get status for pod" podUID="306421d3-af6e-43c8-9411-a8eaaf583862" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.510331 4814 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="c195dca46377ceb3ad0d740239373cf05f82fa6a26d6b06bacf9d304da4b9bfa" exitCode=0 Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.510379 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"c195dca46377ceb3ad0d740239373cf05f82fa6a26d6b06bacf9d304da4b9bfa"} Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.510426 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5426bac6f438de152af29a498c06c72fdd24ef998c5f44f8591b4004b11e7056"} Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.510748 4814 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="929bfc6f-c721-4f45-9674-fc3f151c3f6e" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.510768 4814 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="929bfc6f-c721-4f45-9674-fc3f151c3f6e" Feb 27 16:28:38 crc kubenswrapper[4814]: E0227 16:28:38.511024 4814 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.511078 4814 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.511686 4814 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.512448 4814 status_manager.go:851] "Failed to get status for pod" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" pod="openshift-marketplace/certified-operators-c5wns" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c5wns\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.512966 4814 status_manager.go:851] "Failed to get status for pod" podUID="306421d3-af6e-43c8-9411-a8eaaf583862" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.513386 4814 status_manager.go:851] "Failed to get status for pod" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-nlbk2\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.513736 4814 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.515104 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.516101 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.516184 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e361138fec9135ec90da2b6a0cd44c28a9a5785e42a58404353669569d3afd1d"} Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.517094 4814 status_manager.go:851] "Failed to get status for pod" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-nlbk2\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.517325 4814 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.517478 4814 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.517628 4814 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.517779 4814 status_manager.go:851] "Failed to get status for pod" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" pod="openshift-marketplace/certified-operators-c5wns" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c5wns\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:38 crc kubenswrapper[4814]: I0227 16:28:38.517921 4814 status_manager.go:851] "Failed to get status for pod" podUID="306421d3-af6e-43c8-9411-a8eaaf583862" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.189:6443: connect: connection refused" Feb 27 16:28:39 crc kubenswrapper[4814]: I0227 16:28:39.527494 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4a2c44c8b83b52e16995419b1b23c95f0fb7f08eea828ef3b69481863842af42"} Feb 27 16:28:39 crc kubenswrapper[4814]: I0227 16:28:39.527853 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"62c89140bc2335c3b64ff287df75f42ef133d1721afb27097c9fb5f5d0f43aba"} Feb 27 16:28:40 crc kubenswrapper[4814]: I0227 16:28:40.536786 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"75b595aca9a009eb48c89bfccc276f840c3efc4fbc8b370480b4677d1e84a94b"} Feb 27 16:28:40 crc kubenswrapper[4814]: I0227 16:28:40.536834 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"de52d261e35d79c7ad7108e10fe6c580dad026c285c5ec866bf36fdb4a6a65f7"} Feb 27 16:28:40 crc kubenswrapper[4814]: I0227 16:28:40.536846 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fe483aa7a862eea444481363d4e02a7fc5082e66a21c0ad540555a3764e2815d"} Feb 27 16:28:41 crc kubenswrapper[4814]: I0227 16:28:41.542693 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:41 crc kubenswrapper[4814]: I0227 16:28:41.542724 4814 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="929bfc6f-c721-4f45-9674-fc3f151c3f6e" Feb 27 16:28:41 crc kubenswrapper[4814]: I0227 16:28:41.543141 4814 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="929bfc6f-c721-4f45-9674-fc3f151c3f6e" Feb 27 16:28:42 crc kubenswrapper[4814]: I0227 16:28:42.516800 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:42 crc kubenswrapper[4814]: I0227 16:28:42.516859 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:42 crc kubenswrapper[4814]: I0227 16:28:42.525092 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:43 crc kubenswrapper[4814]: I0227 16:28:43.584572 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:28:43 crc kubenswrapper[4814]: I0227 16:28:43.585133 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:28:43 crc kubenswrapper[4814]: I0227 16:28:43.585177 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:28:43 crc kubenswrapper[4814]: I0227 16:28:43.585247 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:28:43 crc kubenswrapper[4814]: I0227 16:28:43.587975 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 27 16:28:43 crc kubenswrapper[4814]: I0227 16:28:43.588720 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 27 16:28:43 crc kubenswrapper[4814]: I0227 16:28:43.588852 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 27 16:28:43 crc kubenswrapper[4814]: I0227 16:28:43.597162 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 27 16:28:43 crc kubenswrapper[4814]: I0227 16:28:43.606834 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:28:43 crc kubenswrapper[4814]: I0227 16:28:43.614604 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:28:43 crc kubenswrapper[4814]: I0227 16:28:43.615490 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:28:43 crc kubenswrapper[4814]: I0227 16:28:43.616284 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:28:43 crc kubenswrapper[4814]: I0227 16:28:43.669510 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 16:28:43 crc kubenswrapper[4814]: I0227 16:28:43.677461 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:28:43 crc kubenswrapper[4814]: I0227 16:28:43.684751 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 16:28:44 crc kubenswrapper[4814]: W0227 16:28:44.196842 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-37a50949ceed2a3d7f29b24d48bff447b9a0051bef8d84132ccf0313590d06a3 WatchSource:0}: Error finding container 37a50949ceed2a3d7f29b24d48bff447b9a0051bef8d84132ccf0313590d06a3: Status 404 returned error can't find the container with id 37a50949ceed2a3d7f29b24d48bff447b9a0051bef8d84132ccf0313590d06a3 Feb 27 16:28:44 crc kubenswrapper[4814]: W0227 16:28:44.332062 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-a600a6e9f4a458f634970c17e667bf0b6adefb3a888754ae07d0257173dcc7c4 WatchSource:0}: Error finding container a600a6e9f4a458f634970c17e667bf0b6adefb3a888754ae07d0257173dcc7c4: Status 404 returned error can't find the container with id a600a6e9f4a458f634970c17e667bf0b6adefb3a888754ae07d0257173dcc7c4 Feb 27 16:28:44 crc kubenswrapper[4814]: I0227 16:28:44.563841 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"5062c62d6749a532b2bf58a7a25d407161a44fcba8e32fc0c2175527c0a904ed"} Feb 27 16:28:44 crc kubenswrapper[4814]: I0227 16:28:44.563894 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ea550864948dd32c4091163fe6cfbfc3d9fff096df6c61da4f20ce4cf00fbd03"} Feb 27 16:28:44 crc kubenswrapper[4814]: I0227 16:28:44.565900 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"1f6f89b2c96bcd5b450a0e8b2871ca8eda1af9324fd3bd9ce6175ccbe15c7cc8"} Feb 27 16:28:44 crc kubenswrapper[4814]: I0227 16:28:44.565927 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a600a6e9f4a458f634970c17e667bf0b6adefb3a888754ae07d0257173dcc7c4"} Feb 27 16:28:44 crc kubenswrapper[4814]: I0227 16:28:44.566225 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:28:44 crc kubenswrapper[4814]: I0227 16:28:44.567392 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f53001caf6f17d1f8bafaa37f0b268b97bea37b79836a02e9f4b9f46918ca147"} Feb 27 16:28:44 crc kubenswrapper[4814]: I0227 16:28:44.567414 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"37a50949ceed2a3d7f29b24d48bff447b9a0051bef8d84132ccf0313590d06a3"} Feb 27 16:28:46 crc kubenswrapper[4814]: I0227 16:28:46.139311 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" podUID="ec601e1d-d940-4bbb-9cce-83be390511ef" containerName="oauth-openshift" containerID="cri-o://040a0f87d4dd9b296f321af2cb4a9f8f1f9afd14b01d467090cbdde0409be8d8" gracePeriod=15 Feb 27 16:28:46 crc kubenswrapper[4814]: I0227 16:28:46.555100 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:28:46 crc kubenswrapper[4814]: I0227 16:28:46.557811 4814 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:46 crc kubenswrapper[4814]: I0227 16:28:46.562581 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:28:46 crc kubenswrapper[4814]: I0227 16:28:46.584012 4814 generic.go:334] "Generic (PLEG): container finished" podID="ec601e1d-d940-4bbb-9cce-83be390511ef" containerID="040a0f87d4dd9b296f321af2cb4a9f8f1f9afd14b01d467090cbdde0409be8d8" exitCode=0 Feb 27 16:28:46 crc kubenswrapper[4814]: I0227 16:28:46.584197 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" event={"ID":"ec601e1d-d940-4bbb-9cce-83be390511ef","Type":"ContainerDied","Data":"040a0f87d4dd9b296f321af2cb4a9f8f1f9afd14b01d467090cbdde0409be8d8"} Feb 27 16:28:46 crc kubenswrapper[4814]: I0227 16:28:46.585375 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:28:46 crc kubenswrapper[4814]: I0227 16:28:46.600847 4814 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="80c9d16d-5afa-4fa0-88af-b04e681e24ec" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.325236 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.381591 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-serving-cert\") pod \"ec601e1d-d940-4bbb-9cce-83be390511ef\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.381708 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-service-ca\") pod \"ec601e1d-d940-4bbb-9cce-83be390511ef\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.381766 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5v4mm\" (UniqueName: \"kubernetes.io/projected/ec601e1d-d940-4bbb-9cce-83be390511ef-kube-api-access-5v4mm\") pod \"ec601e1d-d940-4bbb-9cce-83be390511ef\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.381828 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-ocp-branding-template\") pod \"ec601e1d-d940-4bbb-9cce-83be390511ef\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.381866 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-idp-0-file-data\") pod \"ec601e1d-d940-4bbb-9cce-83be390511ef\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.381925 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-cliconfig\") pod \"ec601e1d-d940-4bbb-9cce-83be390511ef\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.381999 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ec601e1d-d940-4bbb-9cce-83be390511ef-audit-dir\") pod \"ec601e1d-d940-4bbb-9cce-83be390511ef\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.382054 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-router-certs\") pod \"ec601e1d-d940-4bbb-9cce-83be390511ef\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.382109 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-session\") pod \"ec601e1d-d940-4bbb-9cce-83be390511ef\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.382147 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-login\") pod \"ec601e1d-d940-4bbb-9cce-83be390511ef\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.382186 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-provider-selection\") pod \"ec601e1d-d940-4bbb-9cce-83be390511ef\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.382222 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-audit-policies\") pod \"ec601e1d-d940-4bbb-9cce-83be390511ef\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.382285 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-trusted-ca-bundle\") pod \"ec601e1d-d940-4bbb-9cce-83be390511ef\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.382319 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-error\") pod \"ec601e1d-d940-4bbb-9cce-83be390511ef\" (UID: \"ec601e1d-d940-4bbb-9cce-83be390511ef\") " Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.383831 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ec601e1d-d940-4bbb-9cce-83be390511ef-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "ec601e1d-d940-4bbb-9cce-83be390511ef" (UID: "ec601e1d-d940-4bbb-9cce-83be390511ef"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.384160 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "ec601e1d-d940-4bbb-9cce-83be390511ef" (UID: "ec601e1d-d940-4bbb-9cce-83be390511ef"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.384297 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "ec601e1d-d940-4bbb-9cce-83be390511ef" (UID: "ec601e1d-d940-4bbb-9cce-83be390511ef"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.384404 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "ec601e1d-d940-4bbb-9cce-83be390511ef" (UID: "ec601e1d-d940-4bbb-9cce-83be390511ef"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.384444 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "ec601e1d-d940-4bbb-9cce-83be390511ef" (UID: "ec601e1d-d940-4bbb-9cce-83be390511ef"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.395211 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec601e1d-d940-4bbb-9cce-83be390511ef-kube-api-access-5v4mm" (OuterVolumeSpecName: "kube-api-access-5v4mm") pod "ec601e1d-d940-4bbb-9cce-83be390511ef" (UID: "ec601e1d-d940-4bbb-9cce-83be390511ef"). InnerVolumeSpecName "kube-api-access-5v4mm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.396647 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "ec601e1d-d940-4bbb-9cce-83be390511ef" (UID: "ec601e1d-d940-4bbb-9cce-83be390511ef"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.398859 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "ec601e1d-d940-4bbb-9cce-83be390511ef" (UID: "ec601e1d-d940-4bbb-9cce-83be390511ef"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.399997 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "ec601e1d-d940-4bbb-9cce-83be390511ef" (UID: "ec601e1d-d940-4bbb-9cce-83be390511ef"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.400526 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "ec601e1d-d940-4bbb-9cce-83be390511ef" (UID: "ec601e1d-d940-4bbb-9cce-83be390511ef"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.401632 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "ec601e1d-d940-4bbb-9cce-83be390511ef" (UID: "ec601e1d-d940-4bbb-9cce-83be390511ef"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.401977 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "ec601e1d-d940-4bbb-9cce-83be390511ef" (UID: "ec601e1d-d940-4bbb-9cce-83be390511ef"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.403550 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "ec601e1d-d940-4bbb-9cce-83be390511ef" (UID: "ec601e1d-d940-4bbb-9cce-83be390511ef"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.404878 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "ec601e1d-d940-4bbb-9cce-83be390511ef" (UID: "ec601e1d-d940-4bbb-9cce-83be390511ef"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.484477 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.484524 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.484542 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.484558 4814 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ec601e1d-d940-4bbb-9cce-83be390511ef-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.484574 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.484587 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.484601 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.484616 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.484630 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.484643 4814 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.484656 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.484670 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.484683 4814 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ec601e1d-d940-4bbb-9cce-83be390511ef-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.484696 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5v4mm\" (UniqueName: \"kubernetes.io/projected/ec601e1d-d940-4bbb-9cce-83be390511ef-kube-api-access-5v4mm\") on node \"crc\" DevicePath \"\"" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.595059 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/0.log" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.595149 4814 generic.go:334] "Generic (PLEG): container finished" podID="9d751cbb-f2e2-430d-9754-c882a5e924a5" containerID="5062c62d6749a532b2bf58a7a25d407161a44fcba8e32fc0c2175527c0a904ed" exitCode=255 Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.595285 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerDied","Data":"5062c62d6749a532b2bf58a7a25d407161a44fcba8e32fc0c2175527c0a904ed"} Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.596067 4814 scope.go:117] "RemoveContainer" containerID="5062c62d6749a532b2bf58a7a25d407161a44fcba8e32fc0c2175527c0a904ed" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.599921 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" event={"ID":"ec601e1d-d940-4bbb-9cce-83be390511ef","Type":"ContainerDied","Data":"1f75321f992945c15bb021b573b0ff5038341da7671b291376c15cd482a18aa6"} Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.600000 4814 scope.go:117] "RemoveContainer" containerID="040a0f87d4dd9b296f321af2cb4a9f8f1f9afd14b01d467090cbdde0409be8d8" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.600674 4814 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="929bfc6f-c721-4f45-9674-fc3f151c3f6e" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.600713 4814 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="929bfc6f-c721-4f45-9674-fc3f151c3f6e" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.601918 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-k4p69" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.608233 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.610051 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:28:47 crc kubenswrapper[4814]: I0227 16:28:47.629084 4814 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="80c9d16d-5afa-4fa0-88af-b04e681e24ec" Feb 27 16:28:48 crc kubenswrapper[4814]: I0227 16:28:48.610905 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/0.log" Feb 27 16:28:48 crc kubenswrapper[4814]: I0227 16:28:48.611080 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7948491ccb67cd587f2eb80fd6205ee274d71214b3e6d4b4330e424a0159a394"} Feb 27 16:28:48 crc kubenswrapper[4814]: I0227 16:28:48.614528 4814 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="929bfc6f-c721-4f45-9674-fc3f151c3f6e" Feb 27 16:28:48 crc kubenswrapper[4814]: I0227 16:28:48.614589 4814 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="929bfc6f-c721-4f45-9674-fc3f151c3f6e" Feb 27 16:28:48 crc kubenswrapper[4814]: I0227 16:28:48.637412 4814 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="80c9d16d-5afa-4fa0-88af-b04e681e24ec" Feb 27 16:28:49 crc kubenswrapper[4814]: I0227 16:28:49.624091 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/1.log" Feb 27 16:28:49 crc kubenswrapper[4814]: I0227 16:28:49.625436 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/0.log" Feb 27 16:28:49 crc kubenswrapper[4814]: I0227 16:28:49.625521 4814 generic.go:334] "Generic (PLEG): container finished" podID="9d751cbb-f2e2-430d-9754-c882a5e924a5" containerID="7948491ccb67cd587f2eb80fd6205ee274d71214b3e6d4b4330e424a0159a394" exitCode=255 Feb 27 16:28:49 crc kubenswrapper[4814]: I0227 16:28:49.625572 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerDied","Data":"7948491ccb67cd587f2eb80fd6205ee274d71214b3e6d4b4330e424a0159a394"} Feb 27 16:28:49 crc kubenswrapper[4814]: I0227 16:28:49.625666 4814 scope.go:117] "RemoveContainer" containerID="5062c62d6749a532b2bf58a7a25d407161a44fcba8e32fc0c2175527c0a904ed" Feb 27 16:28:49 crc kubenswrapper[4814]: I0227 16:28:49.626290 4814 scope.go:117] "RemoveContainer" containerID="7948491ccb67cd587f2eb80fd6205ee274d71214b3e6d4b4330e424a0159a394" Feb 27 16:28:49 crc kubenswrapper[4814]: E0227 16:28:49.626532 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=check-endpoints pod=network-check-source-55646444c4-trplf_openshift-network-diagnostics(9d751cbb-f2e2-430d-9754-c882a5e924a5)\"" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:28:50 crc kubenswrapper[4814]: I0227 16:28:50.649248 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/1.log" Feb 27 16:28:54 crc kubenswrapper[4814]: I0227 16:28:54.887701 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 27 16:28:55 crc kubenswrapper[4814]: I0227 16:28:55.496566 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 27 16:28:56 crc kubenswrapper[4814]: I0227 16:28:56.228981 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 27 16:28:56 crc kubenswrapper[4814]: I0227 16:28:56.421388 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 27 16:28:56 crc kubenswrapper[4814]: I0227 16:28:56.593096 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 27 16:28:56 crc kubenswrapper[4814]: I0227 16:28:56.821404 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 27 16:28:56 crc kubenswrapper[4814]: I0227 16:28:56.911769 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 27 16:28:56 crc kubenswrapper[4814]: I0227 16:28:56.917883 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 27 16:28:56 crc kubenswrapper[4814]: I0227 16:28:56.969983 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 27 16:28:57 crc kubenswrapper[4814]: I0227 16:28:57.066882 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 27 16:28:57 crc kubenswrapper[4814]: I0227 16:28:57.460808 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 27 16:28:57 crc kubenswrapper[4814]: I0227 16:28:57.509154 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 27 16:28:57 crc kubenswrapper[4814]: I0227 16:28:57.657586 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 16:28:57 crc kubenswrapper[4814]: I0227 16:28:57.760869 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 27 16:28:57 crc kubenswrapper[4814]: I0227 16:28:57.826644 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 27 16:28:57 crc kubenswrapper[4814]: I0227 16:28:57.830341 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 27 16:28:57 crc kubenswrapper[4814]: I0227 16:28:57.959520 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 27 16:28:58 crc kubenswrapper[4814]: I0227 16:28:58.173115 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 16:28:58 crc kubenswrapper[4814]: I0227 16:28:58.302414 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 27 16:28:58 crc kubenswrapper[4814]: I0227 16:28:58.381581 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 16:28:58 crc kubenswrapper[4814]: I0227 16:28:58.382596 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 27 16:28:58 crc kubenswrapper[4814]: I0227 16:28:58.506144 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 27 16:28:58 crc kubenswrapper[4814]: I0227 16:28:58.565386 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 27 16:28:58 crc kubenswrapper[4814]: I0227 16:28:58.626150 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 27 16:28:58 crc kubenswrapper[4814]: I0227 16:28:58.707360 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 27 16:28:58 crc kubenswrapper[4814]: I0227 16:28:58.731241 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 27 16:28:58 crc kubenswrapper[4814]: I0227 16:28:58.953839 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 27 16:28:59 crc kubenswrapper[4814]: I0227 16:28:59.033794 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 27 16:28:59 crc kubenswrapper[4814]: I0227 16:28:59.166490 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 27 16:28:59 crc kubenswrapper[4814]: I0227 16:28:59.182125 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 27 16:28:59 crc kubenswrapper[4814]: I0227 16:28:59.372491 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 27 16:28:59 crc kubenswrapper[4814]: I0227 16:28:59.424684 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 27 16:28:59 crc kubenswrapper[4814]: I0227 16:28:59.459051 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 27 16:28:59 crc kubenswrapper[4814]: I0227 16:28:59.462103 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 27 16:28:59 crc kubenswrapper[4814]: I0227 16:28:59.474441 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 27 16:28:59 crc kubenswrapper[4814]: I0227 16:28:59.573649 4814 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 27 16:28:59 crc kubenswrapper[4814]: I0227 16:28:59.621929 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 27 16:28:59 crc kubenswrapper[4814]: I0227 16:28:59.677683 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 27 16:28:59 crc kubenswrapper[4814]: I0227 16:28:59.822244 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 27 16:28:59 crc kubenswrapper[4814]: I0227 16:28:59.879113 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 27 16:28:59 crc kubenswrapper[4814]: I0227 16:28:59.905505 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 27 16:28:59 crc kubenswrapper[4814]: I0227 16:28:59.991672 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.020292 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.022328 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.043802 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.063501 4814 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.159552 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.222297 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.280329 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.370278 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.466788 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.542832 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.555808 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.602446 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.644077 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.664465 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.732185 4814 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.741282 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.768812 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.890403 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.928966 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.952072 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.975231 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 27 16:29:00 crc kubenswrapper[4814]: I0227 16:29:00.988697 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.034546 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.081875 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.165174 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.180138 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.212964 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.217918 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.310708 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.389904 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.394085 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.425860 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.489161 4814 scope.go:117] "RemoveContainer" containerID="7948491ccb67cd587f2eb80fd6205ee274d71214b3e6d4b4330e424a0159a394" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.492163 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.504044 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.510543 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.556804 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.579146 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.737876 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/1.log" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.738347 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"6b573b9a7e04d647c80a35f9170caf1f2032ce9bad32eb32b6dbd57c52fc07fb"} Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.760678 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.772399 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.825288 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.855847 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.964642 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 16:29:01 crc kubenswrapper[4814]: I0227 16:29:01.985204 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.019350 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.030915 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.048079 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.270919 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.292044 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.350001 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.448151 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.448481 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.483024 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.515452 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.515772 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.616987 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.734573 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.745818 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/2.log" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.746201 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/1.log" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.746237 4814 generic.go:334] "Generic (PLEG): container finished" podID="9d751cbb-f2e2-430d-9754-c882a5e924a5" containerID="6b573b9a7e04d647c80a35f9170caf1f2032ce9bad32eb32b6dbd57c52fc07fb" exitCode=255 Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.746300 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerDied","Data":"6b573b9a7e04d647c80a35f9170caf1f2032ce9bad32eb32b6dbd57c52fc07fb"} Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.746345 4814 scope.go:117] "RemoveContainer" containerID="7948491ccb67cd587f2eb80fd6205ee274d71214b3e6d4b4330e424a0159a394" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.746880 4814 scope.go:117] "RemoveContainer" containerID="6b573b9a7e04d647c80a35f9170caf1f2032ce9bad32eb32b6dbd57c52fc07fb" Feb 27 16:29:02 crc kubenswrapper[4814]: E0227 16:29:02.747211 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=check-endpoints pod=network-check-source-55646444c4-trplf_openshift-network-diagnostics(9d751cbb-f2e2-430d-9754-c882a5e924a5)\"" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.795973 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.916469 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.969367 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 27 16:29:02 crc kubenswrapper[4814]: I0227 16:29:02.974733 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.017604 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.043189 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.045309 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.129005 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.155598 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.226307 4814 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.231194 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.245770 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.518303 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.688748 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.689676 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.699853 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.756965 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/2.log" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.768244 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.814628 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.827416 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.827518 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.906132 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 27 16:29:03 crc kubenswrapper[4814]: I0227 16:29:03.985672 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 27 16:29:04 crc kubenswrapper[4814]: I0227 16:29:04.040806 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 27 16:29:04 crc kubenswrapper[4814]: I0227 16:29:04.053858 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 27 16:29:04 crc kubenswrapper[4814]: I0227 16:29:04.150574 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 27 16:29:04 crc kubenswrapper[4814]: I0227 16:29:04.333559 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 27 16:29:04 crc kubenswrapper[4814]: I0227 16:29:04.346101 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 27 16:29:04 crc kubenswrapper[4814]: I0227 16:29:04.430481 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 27 16:29:04 crc kubenswrapper[4814]: I0227 16:29:04.464784 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 27 16:29:04 crc kubenswrapper[4814]: I0227 16:29:04.563999 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 27 16:29:04 crc kubenswrapper[4814]: I0227 16:29:04.585917 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 27 16:29:04 crc kubenswrapper[4814]: I0227 16:29:04.703326 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 27 16:29:04 crc kubenswrapper[4814]: I0227 16:29:04.873319 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 27 16:29:04 crc kubenswrapper[4814]: I0227 16:29:04.892851 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 27 16:29:04 crc kubenswrapper[4814]: I0227 16:29:04.935644 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 27 16:29:04 crc kubenswrapper[4814]: I0227 16:29:04.936113 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.052908 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.107965 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.225698 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.309351 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.372989 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.397406 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.421222 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.571922 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.587520 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.609491 4814 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.612725 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=43.61268614 podStartE2EDuration="43.61268614s" podCreationTimestamp="2026-02-27 16:28:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:28:46.620355806 +0000 UTC m=+339.072980666" watchObservedRunningTime="2026-02-27 16:29:05.61268614 +0000 UTC m=+358.065311000" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.627966 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c5wns","openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-k4p69"] Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.628092 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-7db94694d4-47sqk"] Feb 27 16:29:05 crc kubenswrapper[4814]: E0227 16:29:05.628513 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="306421d3-af6e-43c8-9411-a8eaaf583862" containerName="installer" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.628553 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="306421d3-af6e-43c8-9411-a8eaaf583862" containerName="installer" Feb 27 16:29:05 crc kubenswrapper[4814]: E0227 16:29:05.628590 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec601e1d-d940-4bbb-9cce-83be390511ef" containerName="oauth-openshift" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.628604 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec601e1d-d940-4bbb-9cce-83be390511ef" containerName="oauth-openshift" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.628811 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="306421d3-af6e-43c8-9411-a8eaaf583862" containerName="installer" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.628840 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec601e1d-d940-4bbb-9cce-83be390511ef" containerName="oauth-openshift" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.629086 4814 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="929bfc6f-c721-4f45-9674-fc3f151c3f6e" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.629133 4814 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="929bfc6f-c721-4f45-9674-fc3f151c3f6e" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.629649 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.635338 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.635373 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.635430 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.635521 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.636367 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.639462 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.640554 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.640727 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.640912 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.641381 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.643165 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.643566 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.643630 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.646937 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.655551 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.655588 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.659409 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mb6z\" (UniqueName: \"kubernetes.io/projected/28a11698-58dc-4f44-b9e7-194a0b5aab6b-kube-api-access-5mb6z\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.659468 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/28a11698-58dc-4f44-b9e7-194a0b5aab6b-audit-dir\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.659501 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.659589 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-router-certs\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.659638 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.659684 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-session\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.659727 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-user-template-login\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.659799 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.659844 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.659902 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-user-template-error\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.659947 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/28a11698-58dc-4f44-b9e7-194a0b5aab6b-audit-policies\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.660002 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.660036 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.660062 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-service-ca\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.660850 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.704445 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=19.704422399 podStartE2EDuration="19.704422399s" podCreationTimestamp="2026-02-27 16:28:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:29:05.669775634 +0000 UTC m=+358.122400544" watchObservedRunningTime="2026-02-27 16:29:05.704422399 +0000 UTC m=+358.157047239" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.761944 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mb6z\" (UniqueName: \"kubernetes.io/projected/28a11698-58dc-4f44-b9e7-194a0b5aab6b-kube-api-access-5mb6z\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.762110 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/28a11698-58dc-4f44-b9e7-194a0b5aab6b-audit-dir\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.762220 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/28a11698-58dc-4f44-b9e7-194a0b5aab6b-audit-dir\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.762332 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.762377 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-router-certs\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.762412 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.762454 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-session\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.762493 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-user-template-login\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.762573 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.762645 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.762695 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-user-template-error\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.762736 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/28a11698-58dc-4f44-b9e7-194a0b5aab6b-audit-policies\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.762777 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.762819 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.762858 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-service-ca\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.763986 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.764566 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-service-ca\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.765312 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.765469 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/28a11698-58dc-4f44-b9e7-194a0b5aab6b-audit-policies\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.770990 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-router-certs\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.771011 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.771403 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-session\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.773374 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.774247 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.777044 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-user-template-error\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.777480 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-user-template-login\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.778596 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/28a11698-58dc-4f44-b9e7-194a0b5aab6b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.794219 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mb6z\" (UniqueName: \"kubernetes.io/projected/28a11698-58dc-4f44-b9e7-194a0b5aab6b-kube-api-access-5mb6z\") pod \"oauth-openshift-7db94694d4-47sqk\" (UID: \"28a11698-58dc-4f44-b9e7-194a0b5aab6b\") " pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.936014 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.964194 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:05 crc kubenswrapper[4814]: I0227 16:29:05.974475 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.083818 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.171657 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.208790 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.255724 4814 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.294929 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.295567 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.386905 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.404509 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.406351 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.407237 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.501696 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3a647cf-eb96-40b8-982c-8e8b869e57c4" path="/var/lib/kubelet/pods/d3a647cf-eb96-40b8-982c-8e8b869e57c4/volumes" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.502789 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec601e1d-d940-4bbb-9cce-83be390511ef" path="/var/lib/kubelet/pods/ec601e1d-d940-4bbb-9cce-83be390511ef/volumes" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.517084 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7db94694d4-47sqk"] Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.541215 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.551809 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.632170 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.774238 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.785732 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" event={"ID":"28a11698-58dc-4f44-b9e7-194a0b5aab6b","Type":"ContainerStarted","Data":"9571e23bacbf6fe60ae6a5a981bc78ede9904096d62bab1b6226c059b1729121"} Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.794013 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.844458 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.845788 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.894293 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.904401 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.922397 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 27 16:29:06 crc kubenswrapper[4814]: I0227 16:29:06.975907 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.011067 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.094742 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.108840 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.162142 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.300994 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.311626 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.372046 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.408431 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.472824 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.514317 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.521333 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.597747 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.679118 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.689357 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.795947 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" event={"ID":"28a11698-58dc-4f44-b9e7-194a0b5aab6b","Type":"ContainerStarted","Data":"94148b97b259c3ad929784708c5def6a1edc1ca711bd95bc2b6a63a47e1891e8"} Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.796415 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.814116 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.820443 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.838477 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7db94694d4-47sqk" podStartSLOduration=46.838449377 podStartE2EDuration="46.838449377s" podCreationTimestamp="2026-02-27 16:28:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:29:07.836053215 +0000 UTC m=+360.288678085" watchObservedRunningTime="2026-02-27 16:29:07.838449377 +0000 UTC m=+360.291074207" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.851401 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.904156 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 27 16:29:07 crc kubenswrapper[4814]: I0227 16:29:07.962801 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 27 16:29:08 crc kubenswrapper[4814]: I0227 16:29:08.165377 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 27 16:29:08 crc kubenswrapper[4814]: I0227 16:29:08.196902 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 27 16:29:08 crc kubenswrapper[4814]: I0227 16:29:08.262374 4814 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 16:29:08 crc kubenswrapper[4814]: I0227 16:29:08.262777 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://57009b3e63779816d725ceea035b1b0f332445201671efa15bacdd10da79e48a" gracePeriod=5 Feb 27 16:29:08 crc kubenswrapper[4814]: I0227 16:29:08.280866 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 27 16:29:08 crc kubenswrapper[4814]: I0227 16:29:08.351888 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 27 16:29:08 crc kubenswrapper[4814]: I0227 16:29:08.566990 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 27 16:29:08 crc kubenswrapper[4814]: I0227 16:29:08.574821 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 27 16:29:08 crc kubenswrapper[4814]: I0227 16:29:08.646701 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 27 16:29:08 crc kubenswrapper[4814]: I0227 16:29:08.679148 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 27 16:29:08 crc kubenswrapper[4814]: I0227 16:29:08.680940 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 27 16:29:08 crc kubenswrapper[4814]: I0227 16:29:08.698157 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 16:29:08 crc kubenswrapper[4814]: I0227 16:29:08.785409 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 27 16:29:08 crc kubenswrapper[4814]: I0227 16:29:08.848938 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 27 16:29:08 crc kubenswrapper[4814]: I0227 16:29:08.869862 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 27 16:29:08 crc kubenswrapper[4814]: I0227 16:29:08.985638 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 27 16:29:09 crc kubenswrapper[4814]: I0227 16:29:09.011670 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 27 16:29:09 crc kubenswrapper[4814]: I0227 16:29:09.102775 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 27 16:29:09 crc kubenswrapper[4814]: I0227 16:29:09.205208 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 27 16:29:09 crc kubenswrapper[4814]: I0227 16:29:09.348482 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 27 16:29:09 crc kubenswrapper[4814]: I0227 16:29:09.355916 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 27 16:29:09 crc kubenswrapper[4814]: I0227 16:29:09.390886 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 27 16:29:09 crc kubenswrapper[4814]: I0227 16:29:09.516902 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 27 16:29:09 crc kubenswrapper[4814]: I0227 16:29:09.566869 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 27 16:29:09 crc kubenswrapper[4814]: I0227 16:29:09.687982 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 16:29:09 crc kubenswrapper[4814]: I0227 16:29:09.809356 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 27 16:29:09 crc kubenswrapper[4814]: I0227 16:29:09.817357 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 27 16:29:09 crc kubenswrapper[4814]: I0227 16:29:09.849170 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 27 16:29:09 crc kubenswrapper[4814]: I0227 16:29:09.882405 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 27 16:29:09 crc kubenswrapper[4814]: I0227 16:29:09.966934 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 27 16:29:09 crc kubenswrapper[4814]: I0227 16:29:09.970957 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 27 16:29:10 crc kubenswrapper[4814]: I0227 16:29:10.091497 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 27 16:29:10 crc kubenswrapper[4814]: I0227 16:29:10.098037 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 27 16:29:10 crc kubenswrapper[4814]: I0227 16:29:10.434869 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 27 16:29:10 crc kubenswrapper[4814]: I0227 16:29:10.459963 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 16:29:10 crc kubenswrapper[4814]: I0227 16:29:10.708936 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 27 16:29:10 crc kubenswrapper[4814]: I0227 16:29:10.759470 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 27 16:29:10 crc kubenswrapper[4814]: I0227 16:29:10.831177 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 27 16:29:10 crc kubenswrapper[4814]: I0227 16:29:10.902228 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 27 16:29:11 crc kubenswrapper[4814]: I0227 16:29:11.027162 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 27 16:29:11 crc kubenswrapper[4814]: I0227 16:29:11.154524 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 27 16:29:11 crc kubenswrapper[4814]: I0227 16:29:11.249459 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 27 16:29:11 crc kubenswrapper[4814]: I0227 16:29:11.411975 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 27 16:29:11 crc kubenswrapper[4814]: I0227 16:29:11.460050 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 27 16:29:11 crc kubenswrapper[4814]: I0227 16:29:11.588688 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 27 16:29:11 crc kubenswrapper[4814]: I0227 16:29:11.974862 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 27 16:29:13 crc kubenswrapper[4814]: I0227 16:29:13.487954 4814 scope.go:117] "RemoveContainer" containerID="6b573b9a7e04d647c80a35f9170caf1f2032ce9bad32eb32b6dbd57c52fc07fb" Feb 27 16:29:13 crc kubenswrapper[4814]: E0227 16:29:13.488529 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=check-endpoints pod=network-check-source-55646444c4-trplf_openshift-network-diagnostics(9d751cbb-f2e2-430d-9754-c882a5e924a5)\"" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 16:29:13 crc kubenswrapper[4814]: I0227 16:29:13.845958 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 27 16:29:13 crc kubenswrapper[4814]: I0227 16:29:13.846042 4814 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="57009b3e63779816d725ceea035b1b0f332445201671efa15bacdd10da79e48a" exitCode=137 Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.038581 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.038665 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.199566 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.199671 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.199742 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.199781 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.199821 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.199859 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.199919 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.199966 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.200087 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.200336 4814 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.200360 4814 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.200378 4814 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.200394 4814 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.215861 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.302214 4814 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.500603 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.501091 4814 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.517016 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.517067 4814 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="bd6854df-7c7c-46af-87a4-d3d3ea02de16" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.523768 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.523850 4814 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="bd6854df-7c7c-46af-87a4-d3d3ea02de16" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.857361 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.857448 4814 scope.go:117] "RemoveContainer" containerID="57009b3e63779816d725ceea035b1b0f332445201671efa15bacdd10da79e48a" Feb 27 16:29:14 crc kubenswrapper[4814]: I0227 16:29:14.857527 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 16:29:15 crc kubenswrapper[4814]: I0227 16:29:15.809779 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-689bc584d8-f9q99"] Feb 27 16:29:15 crc kubenswrapper[4814]: I0227 16:29:15.810436 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" podUID="3297f6be-fd91-4f82-8d3f-6e1fe627b449" containerName="controller-manager" containerID="cri-o://d52b0abb8f8cbbfa2859761e99c665f2fed9ad8b6ff546737caaee892c4d4731" gracePeriod=30 Feb 27 16:29:15 crc kubenswrapper[4814]: I0227 16:29:15.913948 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz"] Feb 27 16:29:15 crc kubenswrapper[4814]: I0227 16:29:15.914301 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" podUID="10fa0a49-7f2e-4001-a99a-29ec8ce7c921" containerName="route-controller-manager" containerID="cri-o://d70877336baf2a47b41a630c1f36e75d576381e67a5869647a77e9d1f17e9a43" gracePeriod=30 Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.872404 4814 generic.go:334] "Generic (PLEG): container finished" podID="3297f6be-fd91-4f82-8d3f-6e1fe627b449" containerID="d52b0abb8f8cbbfa2859761e99c665f2fed9ad8b6ff546737caaee892c4d4731" exitCode=0 Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.872483 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" event={"ID":"3297f6be-fd91-4f82-8d3f-6e1fe627b449","Type":"ContainerDied","Data":"d52b0abb8f8cbbfa2859761e99c665f2fed9ad8b6ff546737caaee892c4d4731"} Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.872543 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" event={"ID":"3297f6be-fd91-4f82-8d3f-6e1fe627b449","Type":"ContainerDied","Data":"ead29187da7fda139458e1ba01394e20dc3d27538231ed1227798c87069d0751"} Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.872556 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ead29187da7fda139458e1ba01394e20dc3d27538231ed1227798c87069d0751" Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.874410 4814 generic.go:334] "Generic (PLEG): container finished" podID="10fa0a49-7f2e-4001-a99a-29ec8ce7c921" containerID="d70877336baf2a47b41a630c1f36e75d576381e67a5869647a77e9d1f17e9a43" exitCode=0 Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.874521 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" event={"ID":"10fa0a49-7f2e-4001-a99a-29ec8ce7c921","Type":"ContainerDied","Data":"d70877336baf2a47b41a630c1f36e75d576381e67a5869647a77e9d1f17e9a43"} Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.874706 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" event={"ID":"10fa0a49-7f2e-4001-a99a-29ec8ce7c921","Type":"ContainerDied","Data":"ceb5c785cf42ef9597bf9140bb56254a806c04027dad1c0033f079269a54f481"} Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.874727 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ceb5c785cf42ef9597bf9140bb56254a806c04027dad1c0033f079269a54f481" Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.876639 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.895409 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.912453 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-57ff7857bb-qp7x9"] Feb 27 16:29:16 crc kubenswrapper[4814]: E0227 16:29:16.912815 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3297f6be-fd91-4f82-8d3f-6e1fe627b449" containerName="controller-manager" Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.912859 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="3297f6be-fd91-4f82-8d3f-6e1fe627b449" containerName="controller-manager" Feb 27 16:29:16 crc kubenswrapper[4814]: E0227 16:29:16.912876 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10fa0a49-7f2e-4001-a99a-29ec8ce7c921" containerName="route-controller-manager" Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.912885 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="10fa0a49-7f2e-4001-a99a-29ec8ce7c921" containerName="route-controller-manager" Feb 27 16:29:16 crc kubenswrapper[4814]: E0227 16:29:16.912900 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.912907 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.914893 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="3297f6be-fd91-4f82-8d3f-6e1fe627b449" containerName="controller-manager" Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.914935 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.914949 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="10fa0a49-7f2e-4001-a99a-29ec8ce7c921" containerName="route-controller-manager" Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.915769 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:16 crc kubenswrapper[4814]: I0227 16:29:16.926782 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57ff7857bb-qp7x9"] Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.045021 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-client-ca\") pod \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\" (UID: \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\") " Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.045093 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-serving-cert\") pod \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\" (UID: \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\") " Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.045125 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-client-ca\") pod \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.045165 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5f7gh\" (UniqueName: \"kubernetes.io/projected/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-kube-api-access-5f7gh\") pod \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\" (UID: \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\") " Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.045221 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-config\") pod \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\" (UID: \"10fa0a49-7f2e-4001-a99a-29ec8ce7c921\") " Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.045360 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5b278\" (UniqueName: \"kubernetes.io/projected/3297f6be-fd91-4f82-8d3f-6e1fe627b449-kube-api-access-5b278\") pod \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.046215 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-client-ca" (OuterVolumeSpecName: "client-ca") pod "10fa0a49-7f2e-4001-a99a-29ec8ce7c921" (UID: "10fa0a49-7f2e-4001-a99a-29ec8ce7c921"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.045399 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3297f6be-fd91-4f82-8d3f-6e1fe627b449-serving-cert\") pod \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.046650 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-proxy-ca-bundles\") pod \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.046686 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-config\") pod \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\" (UID: \"3297f6be-fd91-4f82-8d3f-6e1fe627b449\") " Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.046951 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d796t\" (UniqueName: \"kubernetes.io/projected/c1a100a6-25f6-4156-a0fd-8d01af0751da-kube-api-access-d796t\") pod \"controller-manager-57ff7857bb-qp7x9\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.046993 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-client-ca\") pod \"controller-manager-57ff7857bb-qp7x9\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.047020 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-proxy-ca-bundles\") pod \"controller-manager-57ff7857bb-qp7x9\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.047051 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-config\") pod \"controller-manager-57ff7857bb-qp7x9\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.047100 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c1a100a6-25f6-4156-a0fd-8d01af0751da-serving-cert\") pod \"controller-manager-57ff7857bb-qp7x9\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.047144 4814 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.047221 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3297f6be-fd91-4f82-8d3f-6e1fe627b449" (UID: "3297f6be-fd91-4f82-8d3f-6e1fe627b449"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.047235 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-config" (OuterVolumeSpecName: "config") pod "10fa0a49-7f2e-4001-a99a-29ec8ce7c921" (UID: "10fa0a49-7f2e-4001-a99a-29ec8ce7c921"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.047239 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-client-ca" (OuterVolumeSpecName: "client-ca") pod "3297f6be-fd91-4f82-8d3f-6e1fe627b449" (UID: "3297f6be-fd91-4f82-8d3f-6e1fe627b449"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.048183 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-config" (OuterVolumeSpecName: "config") pod "3297f6be-fd91-4f82-8d3f-6e1fe627b449" (UID: "3297f6be-fd91-4f82-8d3f-6e1fe627b449"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.051595 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "10fa0a49-7f2e-4001-a99a-29ec8ce7c921" (UID: "10fa0a49-7f2e-4001-a99a-29ec8ce7c921"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.051671 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3297f6be-fd91-4f82-8d3f-6e1fe627b449-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3297f6be-fd91-4f82-8d3f-6e1fe627b449" (UID: "3297f6be-fd91-4f82-8d3f-6e1fe627b449"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.053126 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-kube-api-access-5f7gh" (OuterVolumeSpecName: "kube-api-access-5f7gh") pod "10fa0a49-7f2e-4001-a99a-29ec8ce7c921" (UID: "10fa0a49-7f2e-4001-a99a-29ec8ce7c921"). InnerVolumeSpecName "kube-api-access-5f7gh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.053166 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3297f6be-fd91-4f82-8d3f-6e1fe627b449-kube-api-access-5b278" (OuterVolumeSpecName: "kube-api-access-5b278") pod "3297f6be-fd91-4f82-8d3f-6e1fe627b449" (UID: "3297f6be-fd91-4f82-8d3f-6e1fe627b449"). InnerVolumeSpecName "kube-api-access-5b278". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.148337 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d796t\" (UniqueName: \"kubernetes.io/projected/c1a100a6-25f6-4156-a0fd-8d01af0751da-kube-api-access-d796t\") pod \"controller-manager-57ff7857bb-qp7x9\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.148404 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-client-ca\") pod \"controller-manager-57ff7857bb-qp7x9\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.148436 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-proxy-ca-bundles\") pod \"controller-manager-57ff7857bb-qp7x9\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.148882 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-config\") pod \"controller-manager-57ff7857bb-qp7x9\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.149227 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c1a100a6-25f6-4156-a0fd-8d01af0751da-serving-cert\") pod \"controller-manager-57ff7857bb-qp7x9\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.149466 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5b278\" (UniqueName: \"kubernetes.io/projected/3297f6be-fd91-4f82-8d3f-6e1fe627b449-kube-api-access-5b278\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.149488 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3297f6be-fd91-4f82-8d3f-6e1fe627b449-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.149502 4814 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.149513 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.149526 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.149540 4814 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3297f6be-fd91-4f82-8d3f-6e1fe627b449-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.149554 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5f7gh\" (UniqueName: \"kubernetes.io/projected/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-kube-api-access-5f7gh\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.149566 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10fa0a49-7f2e-4001-a99a-29ec8ce7c921-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.150667 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-proxy-ca-bundles\") pod \"controller-manager-57ff7857bb-qp7x9\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.150959 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-client-ca\") pod \"controller-manager-57ff7857bb-qp7x9\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.151835 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-config\") pod \"controller-manager-57ff7857bb-qp7x9\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.153710 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c1a100a6-25f6-4156-a0fd-8d01af0751da-serving-cert\") pod \"controller-manager-57ff7857bb-qp7x9\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.165899 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d796t\" (UniqueName: \"kubernetes.io/projected/c1a100a6-25f6-4156-a0fd-8d01af0751da-kube-api-access-d796t\") pod \"controller-manager-57ff7857bb-qp7x9\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.252423 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.591682 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v"] Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.592798 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.608906 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v"] Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.696070 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57ff7857bb-qp7x9"] Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.769715 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-config\") pod \"route-controller-manager-8b9b5687d-g276v\" (UID: \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.769896 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-serving-cert\") pod \"route-controller-manager-8b9b5687d-g276v\" (UID: \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.770082 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4fd8\" (UniqueName: \"kubernetes.io/projected/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-kube-api-access-p4fd8\") pod \"route-controller-manager-8b9b5687d-g276v\" (UID: \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.770129 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-client-ca\") pod \"route-controller-manager-8b9b5687d-g276v\" (UID: \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.871959 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4fd8\" (UniqueName: \"kubernetes.io/projected/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-kube-api-access-p4fd8\") pod \"route-controller-manager-8b9b5687d-g276v\" (UID: \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.872031 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-client-ca\") pod \"route-controller-manager-8b9b5687d-g276v\" (UID: \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.872166 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-config\") pod \"route-controller-manager-8b9b5687d-g276v\" (UID: \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.872211 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-serving-cert\") pod \"route-controller-manager-8b9b5687d-g276v\" (UID: \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.873530 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-config\") pod \"route-controller-manager-8b9b5687d-g276v\" (UID: \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.874640 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-client-ca\") pod \"route-controller-manager-8b9b5687d-g276v\" (UID: \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.879835 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-serving-cert\") pod \"route-controller-manager-8b9b5687d-g276v\" (UID: \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.890828 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" event={"ID":"c1a100a6-25f6-4156-a0fd-8d01af0751da","Type":"ContainerStarted","Data":"c73e217c3a49c3aa2d0643298efaea7f877fee2521bb4013ecfe4539bb920deb"} Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.890898 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" event={"ID":"c1a100a6-25f6-4156-a0fd-8d01af0751da","Type":"ContainerStarted","Data":"8bfab45a24a20c1f498429b58836ceedef43e8e80961c42e1efb403b919b97dc"} Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.890897 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-689bc584d8-f9q99" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.890936 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.919974 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4fd8\" (UniqueName: \"kubernetes.io/projected/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-kube-api-access-p4fd8\") pod \"route-controller-manager-8b9b5687d-g276v\" (UID: \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.943057 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-689bc584d8-f9q99"] Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.953318 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-689bc584d8-f9q99"] Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.958029 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz"] Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.964160 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69bc6f769b-bfkvz"] Feb 27 16:29:17 crc kubenswrapper[4814]: I0227 16:29:17.969828 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:18 crc kubenswrapper[4814]: I0227 16:29:18.243168 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v"] Feb 27 16:29:18 crc kubenswrapper[4814]: W0227 16:29:18.248622 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18aa6f43_e27d_45f9_8e7d_16ca9496a44f.slice/crio-9a53319ad86ed04632793dcaab094a865f355574bd390ce7bd55a6dcfc20f1b6 WatchSource:0}: Error finding container 9a53319ad86ed04632793dcaab094a865f355574bd390ce7bd55a6dcfc20f1b6: Status 404 returned error can't find the container with id 9a53319ad86ed04632793dcaab094a865f355574bd390ce7bd55a6dcfc20f1b6 Feb 27 16:29:18 crc kubenswrapper[4814]: I0227 16:29:18.497635 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10fa0a49-7f2e-4001-a99a-29ec8ce7c921" path="/var/lib/kubelet/pods/10fa0a49-7f2e-4001-a99a-29ec8ce7c921/volumes" Feb 27 16:29:18 crc kubenswrapper[4814]: I0227 16:29:18.498852 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3297f6be-fd91-4f82-8d3f-6e1fe627b449" path="/var/lib/kubelet/pods/3297f6be-fd91-4f82-8d3f-6e1fe627b449/volumes" Feb 27 16:29:18 crc kubenswrapper[4814]: I0227 16:29:18.898672 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" event={"ID":"18aa6f43-e27d-45f9-8e7d-16ca9496a44f","Type":"ContainerStarted","Data":"bc08e57ac47d4523e057316b79e34031098a4e919786b59dbf6108c641c153a1"} Feb 27 16:29:18 crc kubenswrapper[4814]: I0227 16:29:18.899077 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:18 crc kubenswrapper[4814]: I0227 16:29:18.899097 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" event={"ID":"18aa6f43-e27d-45f9-8e7d-16ca9496a44f","Type":"ContainerStarted","Data":"9a53319ad86ed04632793dcaab094a865f355574bd390ce7bd55a6dcfc20f1b6"} Feb 27 16:29:18 crc kubenswrapper[4814]: I0227 16:29:18.899850 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:18 crc kubenswrapper[4814]: I0227 16:29:18.905876 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:18 crc kubenswrapper[4814]: I0227 16:29:18.908194 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:18 crc kubenswrapper[4814]: I0227 16:29:18.924180 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" podStartSLOduration=3.924151697 podStartE2EDuration="3.924151697s" podCreationTimestamp="2026-02-27 16:29:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:29:18.91976859 +0000 UTC m=+371.372393420" watchObservedRunningTime="2026-02-27 16:29:18.924151697 +0000 UTC m=+371.376776527" Feb 27 16:29:18 crc kubenswrapper[4814]: I0227 16:29:18.941402 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" podStartSLOduration=3.9413787129999998 podStartE2EDuration="3.941378713s" podCreationTimestamp="2026-02-27 16:29:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:29:18.936643579 +0000 UTC m=+371.389268449" watchObservedRunningTime="2026-02-27 16:29:18.941378713 +0000 UTC m=+371.394003543" Feb 27 16:29:23 crc kubenswrapper[4814]: I0227 16:29:23.685442 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 16:29:24 crc kubenswrapper[4814]: I0227 16:29:24.322033 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 27 16:29:26 crc kubenswrapper[4814]: I0227 16:29:26.487875 4814 scope.go:117] "RemoveContainer" containerID="6b573b9a7e04d647c80a35f9170caf1f2032ce9bad32eb32b6dbd57c52fc07fb" Feb 27 16:29:26 crc kubenswrapper[4814]: I0227 16:29:26.963329 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/2.log" Feb 27 16:29:26 crc kubenswrapper[4814]: I0227 16:29:26.963714 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"3fb059feb0014393addef28ed3d89c30138ff236ea86d8c5f98624cf95d75b60"} Feb 27 16:29:31 crc kubenswrapper[4814]: I0227 16:29:31.011131 4814 generic.go:334] "Generic (PLEG): container finished" podID="dfa49cf2-dc52-4c27-8486-ec1efec61744" containerID="c368bbd9a1ae5371c62ca4d7ecc6b992fe4809aa77ddb9bd665461038649f222" exitCode=0 Feb 27 16:29:31 crc kubenswrapper[4814]: I0227 16:29:31.011293 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" event={"ID":"dfa49cf2-dc52-4c27-8486-ec1efec61744","Type":"ContainerDied","Data":"c368bbd9a1ae5371c62ca4d7ecc6b992fe4809aa77ddb9bd665461038649f222"} Feb 27 16:29:31 crc kubenswrapper[4814]: I0227 16:29:31.012247 4814 scope.go:117] "RemoveContainer" containerID="c368bbd9a1ae5371c62ca4d7ecc6b992fe4809aa77ddb9bd665461038649f222" Feb 27 16:29:32 crc kubenswrapper[4814]: I0227 16:29:32.021807 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" event={"ID":"dfa49cf2-dc52-4c27-8486-ec1efec61744","Type":"ContainerStarted","Data":"fe52ca82eeb498fd270e34ebb923c23f179ca8796e08c87c33acf686f95502db"} Feb 27 16:29:32 crc kubenswrapper[4814]: I0227 16:29:32.022837 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:29:32 crc kubenswrapper[4814]: I0227 16:29:32.026385 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:29:35 crc kubenswrapper[4814]: I0227 16:29:35.810223 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-57ff7857bb-qp7x9"] Feb 27 16:29:35 crc kubenswrapper[4814]: I0227 16:29:35.811144 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" podUID="c1a100a6-25f6-4156-a0fd-8d01af0751da" containerName="controller-manager" containerID="cri-o://c73e217c3a49c3aa2d0643298efaea7f877fee2521bb4013ecfe4539bb920deb" gracePeriod=30 Feb 27 16:29:35 crc kubenswrapper[4814]: I0227 16:29:35.825871 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v"] Feb 27 16:29:35 crc kubenswrapper[4814]: I0227 16:29:35.826229 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" podUID="18aa6f43-e27d-45f9-8e7d-16ca9496a44f" containerName="route-controller-manager" containerID="cri-o://bc08e57ac47d4523e057316b79e34031098a4e919786b59dbf6108c641c153a1" gracePeriod=30 Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.048815 4814 generic.go:334] "Generic (PLEG): container finished" podID="18aa6f43-e27d-45f9-8e7d-16ca9496a44f" containerID="bc08e57ac47d4523e057316b79e34031098a4e919786b59dbf6108c641c153a1" exitCode=0 Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.048915 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" event={"ID":"18aa6f43-e27d-45f9-8e7d-16ca9496a44f","Type":"ContainerDied","Data":"bc08e57ac47d4523e057316b79e34031098a4e919786b59dbf6108c641c153a1"} Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.051409 4814 generic.go:334] "Generic (PLEG): container finished" podID="c1a100a6-25f6-4156-a0fd-8d01af0751da" containerID="c73e217c3a49c3aa2d0643298efaea7f877fee2521bb4013ecfe4539bb920deb" exitCode=0 Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.051455 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" event={"ID":"c1a100a6-25f6-4156-a0fd-8d01af0751da","Type":"ContainerDied","Data":"c73e217c3a49c3aa2d0643298efaea7f877fee2521bb4013ecfe4539bb920deb"} Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.417661 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.560694 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-client-ca\") pod \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\" (UID: \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\") " Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.560828 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-config\") pod \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\" (UID: \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\") " Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.560858 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-serving-cert\") pod \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\" (UID: \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\") " Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.560879 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4fd8\" (UniqueName: \"kubernetes.io/projected/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-kube-api-access-p4fd8\") pod \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\" (UID: \"18aa6f43-e27d-45f9-8e7d-16ca9496a44f\") " Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.561639 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-client-ca" (OuterVolumeSpecName: "client-ca") pod "18aa6f43-e27d-45f9-8e7d-16ca9496a44f" (UID: "18aa6f43-e27d-45f9-8e7d-16ca9496a44f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.561862 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-config" (OuterVolumeSpecName: "config") pod "18aa6f43-e27d-45f9-8e7d-16ca9496a44f" (UID: "18aa6f43-e27d-45f9-8e7d-16ca9496a44f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.566938 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "18aa6f43-e27d-45f9-8e7d-16ca9496a44f" (UID: "18aa6f43-e27d-45f9-8e7d-16ca9496a44f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.574196 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-kube-api-access-p4fd8" (OuterVolumeSpecName: "kube-api-access-p4fd8") pod "18aa6f43-e27d-45f9-8e7d-16ca9496a44f" (UID: "18aa6f43-e27d-45f9-8e7d-16ca9496a44f"). InnerVolumeSpecName "kube-api-access-p4fd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.662928 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.662961 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.662976 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4fd8\" (UniqueName: \"kubernetes.io/projected/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-kube-api-access-p4fd8\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.662986 4814 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18aa6f43-e27d-45f9-8e7d-16ca9496a44f-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.862094 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.967039 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-config\") pod \"c1a100a6-25f6-4156-a0fd-8d01af0751da\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.967188 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d796t\" (UniqueName: \"kubernetes.io/projected/c1a100a6-25f6-4156-a0fd-8d01af0751da-kube-api-access-d796t\") pod \"c1a100a6-25f6-4156-a0fd-8d01af0751da\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.967237 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-client-ca\") pod \"c1a100a6-25f6-4156-a0fd-8d01af0751da\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.967297 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c1a100a6-25f6-4156-a0fd-8d01af0751da-serving-cert\") pod \"c1a100a6-25f6-4156-a0fd-8d01af0751da\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.967336 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-proxy-ca-bundles\") pod \"c1a100a6-25f6-4156-a0fd-8d01af0751da\" (UID: \"c1a100a6-25f6-4156-a0fd-8d01af0751da\") " Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.968119 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c1a100a6-25f6-4156-a0fd-8d01af0751da" (UID: "c1a100a6-25f6-4156-a0fd-8d01af0751da"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.968146 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-client-ca" (OuterVolumeSpecName: "client-ca") pod "c1a100a6-25f6-4156-a0fd-8d01af0751da" (UID: "c1a100a6-25f6-4156-a0fd-8d01af0751da"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.968299 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-config" (OuterVolumeSpecName: "config") pod "c1a100a6-25f6-4156-a0fd-8d01af0751da" (UID: "c1a100a6-25f6-4156-a0fd-8d01af0751da"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.972146 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1a100a6-25f6-4156-a0fd-8d01af0751da-kube-api-access-d796t" (OuterVolumeSpecName: "kube-api-access-d796t") pod "c1a100a6-25f6-4156-a0fd-8d01af0751da" (UID: "c1a100a6-25f6-4156-a0fd-8d01af0751da"). InnerVolumeSpecName "kube-api-access-d796t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:29:36 crc kubenswrapper[4814]: I0227 16:29:36.973322 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1a100a6-25f6-4156-a0fd-8d01af0751da-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c1a100a6-25f6-4156-a0fd-8d01af0751da" (UID: "c1a100a6-25f6-4156-a0fd-8d01af0751da"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.059872 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" event={"ID":"18aa6f43-e27d-45f9-8e7d-16ca9496a44f","Type":"ContainerDied","Data":"9a53319ad86ed04632793dcaab094a865f355574bd390ce7bd55a6dcfc20f1b6"} Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.059917 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.059971 4814 scope.go:117] "RemoveContainer" containerID="bc08e57ac47d4523e057316b79e34031098a4e919786b59dbf6108c641c153a1" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.063893 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" event={"ID":"c1a100a6-25f6-4156-a0fd-8d01af0751da","Type":"ContainerDied","Data":"8bfab45a24a20c1f498429b58836ceedef43e8e80961c42e1efb403b919b97dc"} Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.064075 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57ff7857bb-qp7x9" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.069294 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.069350 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d796t\" (UniqueName: \"kubernetes.io/projected/c1a100a6-25f6-4156-a0fd-8d01af0751da-kube-api-access-d796t\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.069376 4814 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.069404 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c1a100a6-25f6-4156-a0fd-8d01af0751da-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.069424 4814 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c1a100a6-25f6-4156-a0fd-8d01af0751da-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.093216 4814 scope.go:117] "RemoveContainer" containerID="c73e217c3a49c3aa2d0643298efaea7f877fee2521bb4013ecfe4539bb920deb" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.117914 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-57ff7857bb-qp7x9"] Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.127471 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-57ff7857bb-qp7x9"] Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.136115 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v"] Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.144892 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8b9b5687d-g276v"] Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.602530 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5db558bd57-kvkn2"] Feb 27 16:29:37 crc kubenswrapper[4814]: E0227 16:29:37.602830 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18aa6f43-e27d-45f9-8e7d-16ca9496a44f" containerName="route-controller-manager" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.602844 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="18aa6f43-e27d-45f9-8e7d-16ca9496a44f" containerName="route-controller-manager" Feb 27 16:29:37 crc kubenswrapper[4814]: E0227 16:29:37.602862 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1a100a6-25f6-4156-a0fd-8d01af0751da" containerName="controller-manager" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.602870 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1a100a6-25f6-4156-a0fd-8d01af0751da" containerName="controller-manager" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.602991 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="18aa6f43-e27d-45f9-8e7d-16ca9496a44f" containerName="route-controller-manager" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.603010 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1a100a6-25f6-4156-a0fd-8d01af0751da" containerName="controller-manager" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.603511 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.605459 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.606482 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm"] Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.607362 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.607367 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.608898 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.609313 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.611224 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.611430 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.612022 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.612307 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.612358 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.612364 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.612529 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.612722 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.623757 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.626787 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5db558bd57-kvkn2"] Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.650926 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm"] Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.779000 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42bd9983-08bf-4263-8fcb-0714d5067d3a-config\") pod \"route-controller-manager-69c79dd4cc-kmcwm\" (UID: \"42bd9983-08bf-4263-8fcb-0714d5067d3a\") " pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.779445 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-client-ca\") pod \"controller-manager-5db558bd57-kvkn2\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.779765 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbnwl\" (UniqueName: \"kubernetes.io/projected/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-kube-api-access-xbnwl\") pod \"controller-manager-5db558bd57-kvkn2\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.779995 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42bd9983-08bf-4263-8fcb-0714d5067d3a-serving-cert\") pod \"route-controller-manager-69c79dd4cc-kmcwm\" (UID: \"42bd9983-08bf-4263-8fcb-0714d5067d3a\") " pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.780209 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-proxy-ca-bundles\") pod \"controller-manager-5db558bd57-kvkn2\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.780443 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42bd9983-08bf-4263-8fcb-0714d5067d3a-client-ca\") pod \"route-controller-manager-69c79dd4cc-kmcwm\" (UID: \"42bd9983-08bf-4263-8fcb-0714d5067d3a\") " pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.780656 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-config\") pod \"controller-manager-5db558bd57-kvkn2\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.780922 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-serving-cert\") pod \"controller-manager-5db558bd57-kvkn2\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.781130 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfrrl\" (UniqueName: \"kubernetes.io/projected/42bd9983-08bf-4263-8fcb-0714d5067d3a-kube-api-access-gfrrl\") pod \"route-controller-manager-69c79dd4cc-kmcwm\" (UID: \"42bd9983-08bf-4263-8fcb-0714d5067d3a\") " pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.882373 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-serving-cert\") pod \"controller-manager-5db558bd57-kvkn2\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.882735 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfrrl\" (UniqueName: \"kubernetes.io/projected/42bd9983-08bf-4263-8fcb-0714d5067d3a-kube-api-access-gfrrl\") pod \"route-controller-manager-69c79dd4cc-kmcwm\" (UID: \"42bd9983-08bf-4263-8fcb-0714d5067d3a\") " pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.882768 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42bd9983-08bf-4263-8fcb-0714d5067d3a-config\") pod \"route-controller-manager-69c79dd4cc-kmcwm\" (UID: \"42bd9983-08bf-4263-8fcb-0714d5067d3a\") " pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.882796 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-client-ca\") pod \"controller-manager-5db558bd57-kvkn2\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.882822 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbnwl\" (UniqueName: \"kubernetes.io/projected/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-kube-api-access-xbnwl\") pod \"controller-manager-5db558bd57-kvkn2\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.882852 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42bd9983-08bf-4263-8fcb-0714d5067d3a-serving-cert\") pod \"route-controller-manager-69c79dd4cc-kmcwm\" (UID: \"42bd9983-08bf-4263-8fcb-0714d5067d3a\") " pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.882889 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-proxy-ca-bundles\") pod \"controller-manager-5db558bd57-kvkn2\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.882909 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42bd9983-08bf-4263-8fcb-0714d5067d3a-client-ca\") pod \"route-controller-manager-69c79dd4cc-kmcwm\" (UID: \"42bd9983-08bf-4263-8fcb-0714d5067d3a\") " pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.882948 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-config\") pod \"controller-manager-5db558bd57-kvkn2\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.884907 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-config\") pod \"controller-manager-5db558bd57-kvkn2\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.884973 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42bd9983-08bf-4263-8fcb-0714d5067d3a-client-ca\") pod \"route-controller-manager-69c79dd4cc-kmcwm\" (UID: \"42bd9983-08bf-4263-8fcb-0714d5067d3a\") " pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.885098 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-client-ca\") pod \"controller-manager-5db558bd57-kvkn2\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.885925 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42bd9983-08bf-4263-8fcb-0714d5067d3a-config\") pod \"route-controller-manager-69c79dd4cc-kmcwm\" (UID: \"42bd9983-08bf-4263-8fcb-0714d5067d3a\") " pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.886049 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-proxy-ca-bundles\") pod \"controller-manager-5db558bd57-kvkn2\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.889024 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42bd9983-08bf-4263-8fcb-0714d5067d3a-serving-cert\") pod \"route-controller-manager-69c79dd4cc-kmcwm\" (UID: \"42bd9983-08bf-4263-8fcb-0714d5067d3a\") " pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.891390 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-serving-cert\") pod \"controller-manager-5db558bd57-kvkn2\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.904546 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbnwl\" (UniqueName: \"kubernetes.io/projected/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-kube-api-access-xbnwl\") pod \"controller-manager-5db558bd57-kvkn2\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.918243 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfrrl\" (UniqueName: \"kubernetes.io/projected/42bd9983-08bf-4263-8fcb-0714d5067d3a-kube-api-access-gfrrl\") pod \"route-controller-manager-69c79dd4cc-kmcwm\" (UID: \"42bd9983-08bf-4263-8fcb-0714d5067d3a\") " pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.929801 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:37 crc kubenswrapper[4814]: I0227 16:29:37.946913 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:29:38 crc kubenswrapper[4814]: I0227 16:29:38.231598 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm"] Feb 27 16:29:38 crc kubenswrapper[4814]: I0227 16:29:38.375978 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5db558bd57-kvkn2"] Feb 27 16:29:38 crc kubenswrapper[4814]: W0227 16:29:38.381156 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08f14ed2_a40e_47e9_953d_0d2a3cf84a0d.slice/crio-d764b2ae781ddd24491938f625dc7603940bd1810e6a3d7f308d21de92c4f957 WatchSource:0}: Error finding container d764b2ae781ddd24491938f625dc7603940bd1810e6a3d7f308d21de92c4f957: Status 404 returned error can't find the container with id d764b2ae781ddd24491938f625dc7603940bd1810e6a3d7f308d21de92c4f957 Feb 27 16:29:38 crc kubenswrapper[4814]: I0227 16:29:38.504725 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18aa6f43-e27d-45f9-8e7d-16ca9496a44f" path="/var/lib/kubelet/pods/18aa6f43-e27d-45f9-8e7d-16ca9496a44f/volumes" Feb 27 16:29:38 crc kubenswrapper[4814]: I0227 16:29:38.505676 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1a100a6-25f6-4156-a0fd-8d01af0751da" path="/var/lib/kubelet/pods/c1a100a6-25f6-4156-a0fd-8d01af0751da/volumes" Feb 27 16:29:39 crc kubenswrapper[4814]: I0227 16:29:39.078038 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" event={"ID":"42bd9983-08bf-4263-8fcb-0714d5067d3a","Type":"ContainerStarted","Data":"e333cd0383efe72e8735347df435a99911ff844aa5fe15a7dcd0ecf4cd2ea5fd"} Feb 27 16:29:39 crc kubenswrapper[4814]: I0227 16:29:39.078438 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" event={"ID":"42bd9983-08bf-4263-8fcb-0714d5067d3a","Type":"ContainerStarted","Data":"333e221390771e163e4dc47e24e8811f4ba66220e6b61bb663a16a9de9e76486"} Feb 27 16:29:39 crc kubenswrapper[4814]: I0227 16:29:39.078461 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:29:39 crc kubenswrapper[4814]: I0227 16:29:39.080176 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" event={"ID":"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d","Type":"ContainerStarted","Data":"74afe3da5ba4c9243941b03926af4ac2b8cf13e618d045c092c82ffab6cc3ae7"} Feb 27 16:29:39 crc kubenswrapper[4814]: I0227 16:29:39.080243 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" event={"ID":"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d","Type":"ContainerStarted","Data":"d764b2ae781ddd24491938f625dc7603940bd1810e6a3d7f308d21de92c4f957"} Feb 27 16:29:39 crc kubenswrapper[4814]: I0227 16:29:39.080868 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:39 crc kubenswrapper[4814]: I0227 16:29:39.085115 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:29:39 crc kubenswrapper[4814]: I0227 16:29:39.085311 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:29:39 crc kubenswrapper[4814]: I0227 16:29:39.095679 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" podStartSLOduration=4.095666318 podStartE2EDuration="4.095666318s" podCreationTimestamp="2026-02-27 16:29:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:29:39.091750421 +0000 UTC m=+391.544375251" watchObservedRunningTime="2026-02-27 16:29:39.095666318 +0000 UTC m=+391.548291148" Feb 27 16:29:39 crc kubenswrapper[4814]: I0227 16:29:39.134962 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" podStartSLOduration=4.134941794 podStartE2EDuration="4.134941794s" podCreationTimestamp="2026-02-27 16:29:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:29:39.133860089 +0000 UTC m=+391.586484919" watchObservedRunningTime="2026-02-27 16:29:39.134941794 +0000 UTC m=+391.587566624" Feb 27 16:29:42 crc kubenswrapper[4814]: I0227 16:29:42.126317 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 27 16:29:42 crc kubenswrapper[4814]: I0227 16:29:42.220374 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.171952 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536830-xr849"] Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.173217 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536830-xr849" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.176361 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.176359 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.176377 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.183337 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp"] Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.184510 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.186227 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.186361 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.189761 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536830-xr849"] Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.200335 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp"] Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.307169 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtfw6\" (UniqueName: \"kubernetes.io/projected/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-kube-api-access-xtfw6\") pod \"collect-profiles-29536830-6dzbp\" (UID: \"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.307408 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-secret-volume\") pod \"collect-profiles-29536830-6dzbp\" (UID: \"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.307616 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbgnv\" (UniqueName: \"kubernetes.io/projected/c9c67f63-addd-4849-8c2d-3ea2101d7802-kube-api-access-zbgnv\") pod \"auto-csr-approver-29536830-xr849\" (UID: \"c9c67f63-addd-4849-8c2d-3ea2101d7802\") " pod="openshift-infra/auto-csr-approver-29536830-xr849" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.307692 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-config-volume\") pod \"collect-profiles-29536830-6dzbp\" (UID: \"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.409372 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbgnv\" (UniqueName: \"kubernetes.io/projected/c9c67f63-addd-4849-8c2d-3ea2101d7802-kube-api-access-zbgnv\") pod \"auto-csr-approver-29536830-xr849\" (UID: \"c9c67f63-addd-4849-8c2d-3ea2101d7802\") " pod="openshift-infra/auto-csr-approver-29536830-xr849" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.409449 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-config-volume\") pod \"collect-profiles-29536830-6dzbp\" (UID: \"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.409500 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtfw6\" (UniqueName: \"kubernetes.io/projected/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-kube-api-access-xtfw6\") pod \"collect-profiles-29536830-6dzbp\" (UID: \"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.409555 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-secret-volume\") pod \"collect-profiles-29536830-6dzbp\" (UID: \"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.411604 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-config-volume\") pod \"collect-profiles-29536830-6dzbp\" (UID: \"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.415777 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-secret-volume\") pod \"collect-profiles-29536830-6dzbp\" (UID: \"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.439628 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtfw6\" (UniqueName: \"kubernetes.io/projected/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-kube-api-access-xtfw6\") pod \"collect-profiles-29536830-6dzbp\" (UID: \"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.441862 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbgnv\" (UniqueName: \"kubernetes.io/projected/c9c67f63-addd-4849-8c2d-3ea2101d7802-kube-api-access-zbgnv\") pod \"auto-csr-approver-29536830-xr849\" (UID: \"c9c67f63-addd-4849-8c2d-3ea2101d7802\") " pod="openshift-infra/auto-csr-approver-29536830-xr849" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.503425 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536830-xr849" Feb 27 16:30:00 crc kubenswrapper[4814]: I0227 16:30:00.522439 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp" Feb 27 16:30:01 crc kubenswrapper[4814]: I0227 16:30:01.028390 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536830-xr849"] Feb 27 16:30:01 crc kubenswrapper[4814]: W0227 16:30:01.034692 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9c67f63_addd_4849_8c2d_3ea2101d7802.slice/crio-a6a8ac307074010c88a39d4f7987bfe2a8b0c1763a2ca0e8fd7a4a2b616a69d3 WatchSource:0}: Error finding container a6a8ac307074010c88a39d4f7987bfe2a8b0c1763a2ca0e8fd7a4a2b616a69d3: Status 404 returned error can't find the container with id a6a8ac307074010c88a39d4f7987bfe2a8b0c1763a2ca0e8fd7a4a2b616a69d3 Feb 27 16:30:01 crc kubenswrapper[4814]: I0227 16:30:01.113805 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp"] Feb 27 16:30:01 crc kubenswrapper[4814]: W0227 16:30:01.119712 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a75f56f_a4b8_413c_8bc0_b56f3eea4cc6.slice/crio-149bb2fd57661bbc780e114b092dee4181f0f4ea4487bfe38d12239e2ce28ddf WatchSource:0}: Error finding container 149bb2fd57661bbc780e114b092dee4181f0f4ea4487bfe38d12239e2ce28ddf: Status 404 returned error can't find the container with id 149bb2fd57661bbc780e114b092dee4181f0f4ea4487bfe38d12239e2ce28ddf Feb 27 16:30:01 crc kubenswrapper[4814]: I0227 16:30:01.220158 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536830-xr849" event={"ID":"c9c67f63-addd-4849-8c2d-3ea2101d7802","Type":"ContainerStarted","Data":"a6a8ac307074010c88a39d4f7987bfe2a8b0c1763a2ca0e8fd7a4a2b616a69d3"} Feb 27 16:30:01 crc kubenswrapper[4814]: I0227 16:30:01.223609 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp" event={"ID":"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6","Type":"ContainerStarted","Data":"149bb2fd57661bbc780e114b092dee4181f0f4ea4487bfe38d12239e2ce28ddf"} Feb 27 16:30:02 crc kubenswrapper[4814]: I0227 16:30:02.237694 4814 generic.go:334] "Generic (PLEG): container finished" podID="1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6" containerID="2dfa94b8890e8410ea578e0769f6b244b7eda1951bdf77e6e572a5b0c8b509fc" exitCode=0 Feb 27 16:30:02 crc kubenswrapper[4814]: I0227 16:30:02.237747 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp" event={"ID":"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6","Type":"ContainerDied","Data":"2dfa94b8890e8410ea578e0769f6b244b7eda1951bdf77e6e572a5b0c8b509fc"} Feb 27 16:30:03 crc kubenswrapper[4814]: I0227 16:30:03.245204 4814 generic.go:334] "Generic (PLEG): container finished" podID="c9c67f63-addd-4849-8c2d-3ea2101d7802" containerID="f54787b0135ae6183bcf18cdd7f2beef6bd7b6d884e1079dd611efbdfe04e5c8" exitCode=0 Feb 27 16:30:03 crc kubenswrapper[4814]: I0227 16:30:03.245656 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536830-xr849" event={"ID":"c9c67f63-addd-4849-8c2d-3ea2101d7802","Type":"ContainerDied","Data":"f54787b0135ae6183bcf18cdd7f2beef6bd7b6d884e1079dd611efbdfe04e5c8"} Feb 27 16:30:03 crc kubenswrapper[4814]: I0227 16:30:03.662085 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp" Feb 27 16:30:03 crc kubenswrapper[4814]: I0227 16:30:03.757905 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-config-volume\") pod \"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6\" (UID: \"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6\") " Feb 27 16:30:03 crc kubenswrapper[4814]: I0227 16:30:03.758041 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtfw6\" (UniqueName: \"kubernetes.io/projected/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-kube-api-access-xtfw6\") pod \"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6\" (UID: \"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6\") " Feb 27 16:30:03 crc kubenswrapper[4814]: I0227 16:30:03.758093 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-secret-volume\") pod \"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6\" (UID: \"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6\") " Feb 27 16:30:03 crc kubenswrapper[4814]: I0227 16:30:03.759804 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-config-volume" (OuterVolumeSpecName: "config-volume") pod "1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6" (UID: "1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:30:03 crc kubenswrapper[4814]: I0227 16:30:03.762987 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6" (UID: "1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:30:03 crc kubenswrapper[4814]: I0227 16:30:03.764338 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-kube-api-access-xtfw6" (OuterVolumeSpecName: "kube-api-access-xtfw6") pod "1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6" (UID: "1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6"). InnerVolumeSpecName "kube-api-access-xtfw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:30:03 crc kubenswrapper[4814]: I0227 16:30:03.859194 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtfw6\" (UniqueName: \"kubernetes.io/projected/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-kube-api-access-xtfw6\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:03 crc kubenswrapper[4814]: I0227 16:30:03.859287 4814 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:03 crc kubenswrapper[4814]: I0227 16:30:03.859316 4814 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:04 crc kubenswrapper[4814]: I0227 16:30:04.259486 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp" Feb 27 16:30:04 crc kubenswrapper[4814]: I0227 16:30:04.259492 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp" event={"ID":"1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6","Type":"ContainerDied","Data":"149bb2fd57661bbc780e114b092dee4181f0f4ea4487bfe38d12239e2ce28ddf"} Feb 27 16:30:04 crc kubenswrapper[4814]: I0227 16:30:04.259605 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="149bb2fd57661bbc780e114b092dee4181f0f4ea4487bfe38d12239e2ce28ddf" Feb 27 16:30:04 crc kubenswrapper[4814]: I0227 16:30:04.737272 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536830-xr849" Feb 27 16:30:04 crc kubenswrapper[4814]: I0227 16:30:04.877431 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbgnv\" (UniqueName: \"kubernetes.io/projected/c9c67f63-addd-4849-8c2d-3ea2101d7802-kube-api-access-zbgnv\") pod \"c9c67f63-addd-4849-8c2d-3ea2101d7802\" (UID: \"c9c67f63-addd-4849-8c2d-3ea2101d7802\") " Feb 27 16:30:04 crc kubenswrapper[4814]: I0227 16:30:04.883351 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9c67f63-addd-4849-8c2d-3ea2101d7802-kube-api-access-zbgnv" (OuterVolumeSpecName: "kube-api-access-zbgnv") pod "c9c67f63-addd-4849-8c2d-3ea2101d7802" (UID: "c9c67f63-addd-4849-8c2d-3ea2101d7802"). InnerVolumeSpecName "kube-api-access-zbgnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:30:04 crc kubenswrapper[4814]: I0227 16:30:04.979328 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbgnv\" (UniqueName: \"kubernetes.io/projected/c9c67f63-addd-4849-8c2d-3ea2101d7802-kube-api-access-zbgnv\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:05 crc kubenswrapper[4814]: I0227 16:30:05.270366 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536830-xr849" event={"ID":"c9c67f63-addd-4849-8c2d-3ea2101d7802","Type":"ContainerDied","Data":"a6a8ac307074010c88a39d4f7987bfe2a8b0c1763a2ca0e8fd7a4a2b616a69d3"} Feb 27 16:30:05 crc kubenswrapper[4814]: I0227 16:30:05.270431 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6a8ac307074010c88a39d4f7987bfe2a8b0c1763a2ca0e8fd7a4a2b616a69d3" Feb 27 16:30:05 crc kubenswrapper[4814]: I0227 16:30:05.270444 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536830-xr849" Feb 27 16:30:15 crc kubenswrapper[4814]: I0227 16:30:15.823573 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5db558bd57-kvkn2"] Feb 27 16:30:15 crc kubenswrapper[4814]: I0227 16:30:15.825808 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" podUID="08f14ed2-a40e-47e9-953d-0d2a3cf84a0d" containerName="controller-manager" containerID="cri-o://74afe3da5ba4c9243941b03926af4ac2b8cf13e618d045c092c82ffab6cc3ae7" gracePeriod=30 Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.366548 4814 generic.go:334] "Generic (PLEG): container finished" podID="08f14ed2-a40e-47e9-953d-0d2a3cf84a0d" containerID="74afe3da5ba4c9243941b03926af4ac2b8cf13e618d045c092c82ffab6cc3ae7" exitCode=0 Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.366646 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" event={"ID":"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d","Type":"ContainerDied","Data":"74afe3da5ba4c9243941b03926af4ac2b8cf13e618d045c092c82ffab6cc3ae7"} Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.367058 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" event={"ID":"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d","Type":"ContainerDied","Data":"d764b2ae781ddd24491938f625dc7603940bd1810e6a3d7f308d21de92c4f957"} Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.367077 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d764b2ae781ddd24491938f625dc7603940bd1810e6a3d7f308d21de92c4f957" Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.395968 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.459415 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbnwl\" (UniqueName: \"kubernetes.io/projected/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-kube-api-access-xbnwl\") pod \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.459508 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-client-ca\") pod \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.459571 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-config\") pod \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.459618 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-proxy-ca-bundles\") pod \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.459651 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-serving-cert\") pod \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\" (UID: \"08f14ed2-a40e-47e9-953d-0d2a3cf84a0d\") " Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.460931 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "08f14ed2-a40e-47e9-953d-0d2a3cf84a0d" (UID: "08f14ed2-a40e-47e9-953d-0d2a3cf84a0d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.460950 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-config" (OuterVolumeSpecName: "config") pod "08f14ed2-a40e-47e9-953d-0d2a3cf84a0d" (UID: "08f14ed2-a40e-47e9-953d-0d2a3cf84a0d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.461106 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-client-ca" (OuterVolumeSpecName: "client-ca") pod "08f14ed2-a40e-47e9-953d-0d2a3cf84a0d" (UID: "08f14ed2-a40e-47e9-953d-0d2a3cf84a0d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.467611 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "08f14ed2-a40e-47e9-953d-0d2a3cf84a0d" (UID: "08f14ed2-a40e-47e9-953d-0d2a3cf84a0d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.468467 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-kube-api-access-xbnwl" (OuterVolumeSpecName: "kube-api-access-xbnwl") pod "08f14ed2-a40e-47e9-953d-0d2a3cf84a0d" (UID: "08f14ed2-a40e-47e9-953d-0d2a3cf84a0d"). InnerVolumeSpecName "kube-api-access-xbnwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.561520 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbnwl\" (UniqueName: \"kubernetes.io/projected/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-kube-api-access-xbnwl\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.561632 4814 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.561650 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.561664 4814 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:16 crc kubenswrapper[4814]: I0227 16:30:16.561683 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.374716 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5db558bd57-kvkn2" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.396793 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5db558bd57-kvkn2"] Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.401314 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5db558bd57-kvkn2"] Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.641167 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-57ff7857bb-ftwbb"] Feb 27 16:30:17 crc kubenswrapper[4814]: E0227 16:30:17.642151 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9c67f63-addd-4849-8c2d-3ea2101d7802" containerName="oc" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.642198 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9c67f63-addd-4849-8c2d-3ea2101d7802" containerName="oc" Feb 27 16:30:17 crc kubenswrapper[4814]: E0227 16:30:17.642293 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6" containerName="collect-profiles" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.642315 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6" containerName="collect-profiles" Feb 27 16:30:17 crc kubenswrapper[4814]: E0227 16:30:17.642359 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08f14ed2-a40e-47e9-953d-0d2a3cf84a0d" containerName="controller-manager" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.642379 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="08f14ed2-a40e-47e9-953d-0d2a3cf84a0d" containerName="controller-manager" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.642893 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="08f14ed2-a40e-47e9-953d-0d2a3cf84a0d" containerName="controller-manager" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.642981 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9c67f63-addd-4849-8c2d-3ea2101d7802" containerName="oc" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.643017 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6" containerName="collect-profiles" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.644242 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.653030 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.653603 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.654017 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.654918 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.654965 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.655896 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.664964 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57ff7857bb-ftwbb"] Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.668788 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.678124 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/254da4ce-101b-4a32-a8e9-1f9c0a2d5699-client-ca\") pod \"controller-manager-57ff7857bb-ftwbb\" (UID: \"254da4ce-101b-4a32-a8e9-1f9c0a2d5699\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.678209 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/254da4ce-101b-4a32-a8e9-1f9c0a2d5699-serving-cert\") pod \"controller-manager-57ff7857bb-ftwbb\" (UID: \"254da4ce-101b-4a32-a8e9-1f9c0a2d5699\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.678296 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/254da4ce-101b-4a32-a8e9-1f9c0a2d5699-config\") pod \"controller-manager-57ff7857bb-ftwbb\" (UID: \"254da4ce-101b-4a32-a8e9-1f9c0a2d5699\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.678330 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkddk\" (UniqueName: \"kubernetes.io/projected/254da4ce-101b-4a32-a8e9-1f9c0a2d5699-kube-api-access-hkddk\") pod \"controller-manager-57ff7857bb-ftwbb\" (UID: \"254da4ce-101b-4a32-a8e9-1f9c0a2d5699\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.678384 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/254da4ce-101b-4a32-a8e9-1f9c0a2d5699-proxy-ca-bundles\") pod \"controller-manager-57ff7857bb-ftwbb\" (UID: \"254da4ce-101b-4a32-a8e9-1f9c0a2d5699\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.779542 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/254da4ce-101b-4a32-a8e9-1f9c0a2d5699-config\") pod \"controller-manager-57ff7857bb-ftwbb\" (UID: \"254da4ce-101b-4a32-a8e9-1f9c0a2d5699\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.779658 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkddk\" (UniqueName: \"kubernetes.io/projected/254da4ce-101b-4a32-a8e9-1f9c0a2d5699-kube-api-access-hkddk\") pod \"controller-manager-57ff7857bb-ftwbb\" (UID: \"254da4ce-101b-4a32-a8e9-1f9c0a2d5699\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.779711 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/254da4ce-101b-4a32-a8e9-1f9c0a2d5699-proxy-ca-bundles\") pod \"controller-manager-57ff7857bb-ftwbb\" (UID: \"254da4ce-101b-4a32-a8e9-1f9c0a2d5699\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.779769 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/254da4ce-101b-4a32-a8e9-1f9c0a2d5699-client-ca\") pod \"controller-manager-57ff7857bb-ftwbb\" (UID: \"254da4ce-101b-4a32-a8e9-1f9c0a2d5699\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.779831 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/254da4ce-101b-4a32-a8e9-1f9c0a2d5699-serving-cert\") pod \"controller-manager-57ff7857bb-ftwbb\" (UID: \"254da4ce-101b-4a32-a8e9-1f9c0a2d5699\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.781432 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/254da4ce-101b-4a32-a8e9-1f9c0a2d5699-proxy-ca-bundles\") pod \"controller-manager-57ff7857bb-ftwbb\" (UID: \"254da4ce-101b-4a32-a8e9-1f9c0a2d5699\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.781594 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/254da4ce-101b-4a32-a8e9-1f9c0a2d5699-client-ca\") pod \"controller-manager-57ff7857bb-ftwbb\" (UID: \"254da4ce-101b-4a32-a8e9-1f9c0a2d5699\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.783876 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/254da4ce-101b-4a32-a8e9-1f9c0a2d5699-config\") pod \"controller-manager-57ff7857bb-ftwbb\" (UID: \"254da4ce-101b-4a32-a8e9-1f9c0a2d5699\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.787698 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/254da4ce-101b-4a32-a8e9-1f9c0a2d5699-serving-cert\") pod \"controller-manager-57ff7857bb-ftwbb\" (UID: \"254da4ce-101b-4a32-a8e9-1f9c0a2d5699\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.810606 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkddk\" (UniqueName: \"kubernetes.io/projected/254da4ce-101b-4a32-a8e9-1f9c0a2d5699-kube-api-access-hkddk\") pod \"controller-manager-57ff7857bb-ftwbb\" (UID: \"254da4ce-101b-4a32-a8e9-1f9c0a2d5699\") " pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:17 crc kubenswrapper[4814]: I0227 16:30:17.977951 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:18 crc kubenswrapper[4814]: I0227 16:30:18.507414 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08f14ed2-a40e-47e9-953d-0d2a3cf84a0d" path="/var/lib/kubelet/pods/08f14ed2-a40e-47e9-953d-0d2a3cf84a0d/volumes" Feb 27 16:30:18 crc kubenswrapper[4814]: I0227 16:30:18.642535 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57ff7857bb-ftwbb"] Feb 27 16:30:18 crc kubenswrapper[4814]: W0227 16:30:18.650034 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod254da4ce_101b_4a32_a8e9_1f9c0a2d5699.slice/crio-363118a7e8b7acb635b438002eb0a1a3599b02432b29b1483d6a4b72b7785e79 WatchSource:0}: Error finding container 363118a7e8b7acb635b438002eb0a1a3599b02432b29b1483d6a4b72b7785e79: Status 404 returned error can't find the container with id 363118a7e8b7acb635b438002eb0a1a3599b02432b29b1483d6a4b72b7785e79 Feb 27 16:30:19 crc kubenswrapper[4814]: I0227 16:30:19.389105 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" event={"ID":"254da4ce-101b-4a32-a8e9-1f9c0a2d5699","Type":"ContainerStarted","Data":"d7d37339173ee25f6add43e5d50e5267fec17df99bc6bb97163b6901e84a6286"} Feb 27 16:30:19 crc kubenswrapper[4814]: I0227 16:30:19.389751 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:19 crc kubenswrapper[4814]: I0227 16:30:19.389768 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" event={"ID":"254da4ce-101b-4a32-a8e9-1f9c0a2d5699","Type":"ContainerStarted","Data":"363118a7e8b7acb635b438002eb0a1a3599b02432b29b1483d6a4b72b7785e79"} Feb 27 16:30:19 crc kubenswrapper[4814]: I0227 16:30:19.394414 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" Feb 27 16:30:19 crc kubenswrapper[4814]: I0227 16:30:19.439559 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-57ff7857bb-ftwbb" podStartSLOduration=4.439537156 podStartE2EDuration="4.439537156s" podCreationTimestamp="2026-02-27 16:30:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:30:19.41376886 +0000 UTC m=+431.866393690" watchObservedRunningTime="2026-02-27 16:30:19.439537156 +0000 UTC m=+431.892161986" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.373649 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wj4c9"] Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.375376 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.401036 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wj4c9"] Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.459517 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.459952 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-trusted-ca\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.460153 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-bound-sa-token\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.460440 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsbqr\" (UniqueName: \"kubernetes.io/projected/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-kube-api-access-bsbqr\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.460749 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-registry-tls\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.460958 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-registry-certificates\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.461165 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.461383 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.487764 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.562382 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.562428 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.562464 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-trusted-ca\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.562493 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-bound-sa-token\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.562531 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsbqr\" (UniqueName: \"kubernetes.io/projected/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-kube-api-access-bsbqr\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.562573 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-registry-tls\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.562594 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-registry-certificates\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.563061 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.564354 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-registry-certificates\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.564849 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-trusted-ca\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.576285 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-registry-tls\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.579973 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-bound-sa-token\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.584371 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsbqr\" (UniqueName: \"kubernetes.io/projected/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-kube-api-access-bsbqr\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.588042 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7bed7a8d-18b8-48bf-a436-d8e8c55cebcb-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wj4c9\" (UID: \"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb\") " pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:28 crc kubenswrapper[4814]: I0227 16:30:28.697081 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:29 crc kubenswrapper[4814]: I0227 16:30:29.180546 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wj4c9"] Feb 27 16:30:29 crc kubenswrapper[4814]: I0227 16:30:29.485031 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" event={"ID":"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb","Type":"ContainerStarted","Data":"868459a8ec0702baef715a19a3438bd337fbb9e2b44e8871ba94d6148dcf827f"} Feb 27 16:30:29 crc kubenswrapper[4814]: I0227 16:30:29.485981 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:29 crc kubenswrapper[4814]: I0227 16:30:29.486175 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" event={"ID":"7bed7a8d-18b8-48bf-a436-d8e8c55cebcb","Type":"ContainerStarted","Data":"61d0db1f51b639f3bc8806219eabf6f43c978a7cad978b157a761c3c7e4aa661"} Feb 27 16:30:29 crc kubenswrapper[4814]: I0227 16:30:29.524723 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" podStartSLOduration=1.5246991319999998 podStartE2EDuration="1.524699132s" podCreationTimestamp="2026-02-27 16:30:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:30:29.523439642 +0000 UTC m=+441.976064482" watchObservedRunningTime="2026-02-27 16:30:29.524699132 +0000 UTC m=+441.977323962" Feb 27 16:30:35 crc kubenswrapper[4814]: I0227 16:30:35.827724 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm"] Feb 27 16:30:35 crc kubenswrapper[4814]: I0227 16:30:35.828424 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" podUID="42bd9983-08bf-4263-8fcb-0714d5067d3a" containerName="route-controller-manager" containerID="cri-o://e333cd0383efe72e8735347df435a99911ff844aa5fe15a7dcd0ecf4cd2ea5fd" gracePeriod=30 Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.449766 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.535866 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42bd9983-08bf-4263-8fcb-0714d5067d3a-config\") pod \"42bd9983-08bf-4263-8fcb-0714d5067d3a\" (UID: \"42bd9983-08bf-4263-8fcb-0714d5067d3a\") " Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.536041 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42bd9983-08bf-4263-8fcb-0714d5067d3a-client-ca\") pod \"42bd9983-08bf-4263-8fcb-0714d5067d3a\" (UID: \"42bd9983-08bf-4263-8fcb-0714d5067d3a\") " Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.536095 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfrrl\" (UniqueName: \"kubernetes.io/projected/42bd9983-08bf-4263-8fcb-0714d5067d3a-kube-api-access-gfrrl\") pod \"42bd9983-08bf-4263-8fcb-0714d5067d3a\" (UID: \"42bd9983-08bf-4263-8fcb-0714d5067d3a\") " Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.536156 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42bd9983-08bf-4263-8fcb-0714d5067d3a-serving-cert\") pod \"42bd9983-08bf-4263-8fcb-0714d5067d3a\" (UID: \"42bd9983-08bf-4263-8fcb-0714d5067d3a\") " Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.538760 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42bd9983-08bf-4263-8fcb-0714d5067d3a-client-ca" (OuterVolumeSpecName: "client-ca") pod "42bd9983-08bf-4263-8fcb-0714d5067d3a" (UID: "42bd9983-08bf-4263-8fcb-0714d5067d3a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.539523 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42bd9983-08bf-4263-8fcb-0714d5067d3a-config" (OuterVolumeSpecName: "config") pod "42bd9983-08bf-4263-8fcb-0714d5067d3a" (UID: "42bd9983-08bf-4263-8fcb-0714d5067d3a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.546538 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42bd9983-08bf-4263-8fcb-0714d5067d3a-kube-api-access-gfrrl" (OuterVolumeSpecName: "kube-api-access-gfrrl") pod "42bd9983-08bf-4263-8fcb-0714d5067d3a" (UID: "42bd9983-08bf-4263-8fcb-0714d5067d3a"). InnerVolumeSpecName "kube-api-access-gfrrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.548555 4814 generic.go:334] "Generic (PLEG): container finished" podID="42bd9983-08bf-4263-8fcb-0714d5067d3a" containerID="e333cd0383efe72e8735347df435a99911ff844aa5fe15a7dcd0ecf4cd2ea5fd" exitCode=0 Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.548609 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" event={"ID":"42bd9983-08bf-4263-8fcb-0714d5067d3a","Type":"ContainerDied","Data":"e333cd0383efe72e8735347df435a99911ff844aa5fe15a7dcd0ecf4cd2ea5fd"} Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.548625 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.548646 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm" event={"ID":"42bd9983-08bf-4263-8fcb-0714d5067d3a","Type":"ContainerDied","Data":"333e221390771e163e4dc47e24e8811f4ba66220e6b61bb663a16a9de9e76486"} Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.548668 4814 scope.go:117] "RemoveContainer" containerID="e333cd0383efe72e8735347df435a99911ff844aa5fe15a7dcd0ecf4cd2ea5fd" Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.560748 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42bd9983-08bf-4263-8fcb-0714d5067d3a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "42bd9983-08bf-4263-8fcb-0714d5067d3a" (UID: "42bd9983-08bf-4263-8fcb-0714d5067d3a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.599495 4814 scope.go:117] "RemoveContainer" containerID="e333cd0383efe72e8735347df435a99911ff844aa5fe15a7dcd0ecf4cd2ea5fd" Feb 27 16:30:36 crc kubenswrapper[4814]: E0227 16:30:36.600347 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e333cd0383efe72e8735347df435a99911ff844aa5fe15a7dcd0ecf4cd2ea5fd\": container with ID starting with e333cd0383efe72e8735347df435a99911ff844aa5fe15a7dcd0ecf4cd2ea5fd not found: ID does not exist" containerID="e333cd0383efe72e8735347df435a99911ff844aa5fe15a7dcd0ecf4cd2ea5fd" Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.600556 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e333cd0383efe72e8735347df435a99911ff844aa5fe15a7dcd0ecf4cd2ea5fd"} err="failed to get container status \"e333cd0383efe72e8735347df435a99911ff844aa5fe15a7dcd0ecf4cd2ea5fd\": rpc error: code = NotFound desc = could not find container \"e333cd0383efe72e8735347df435a99911ff844aa5fe15a7dcd0ecf4cd2ea5fd\": container with ID starting with e333cd0383efe72e8735347df435a99911ff844aa5fe15a7dcd0ecf4cd2ea5fd not found: ID does not exist" Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.639494 4814 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42bd9983-08bf-4263-8fcb-0714d5067d3a-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.639885 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfrrl\" (UniqueName: \"kubernetes.io/projected/42bd9983-08bf-4263-8fcb-0714d5067d3a-kube-api-access-gfrrl\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.640043 4814 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42bd9983-08bf-4263-8fcb-0714d5067d3a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.640168 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42bd9983-08bf-4263-8fcb-0714d5067d3a-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.905945 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm"] Feb 27 16:30:36 crc kubenswrapper[4814]: I0227 16:30:36.909893 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69c79dd4cc-kmcwm"] Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.655599 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r"] Feb 27 16:30:37 crc kubenswrapper[4814]: E0227 16:30:37.656436 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42bd9983-08bf-4263-8fcb-0714d5067d3a" containerName="route-controller-manager" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.656480 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="42bd9983-08bf-4263-8fcb-0714d5067d3a" containerName="route-controller-manager" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.656740 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="42bd9983-08bf-4263-8fcb-0714d5067d3a" containerName="route-controller-manager" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.657404 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.659815 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.662623 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.663306 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.665042 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.665202 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.665374 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.672677 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r"] Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.758328 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92cf2506-fc0d-4227-9195-7daa9d044919-serving-cert\") pod \"route-controller-manager-8b9b5687d-wkb5r\" (UID: \"92cf2506-fc0d-4227-9195-7daa9d044919\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.758367 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92cf2506-fc0d-4227-9195-7daa9d044919-client-ca\") pod \"route-controller-manager-8b9b5687d-wkb5r\" (UID: \"92cf2506-fc0d-4227-9195-7daa9d044919\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.758396 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92cf2506-fc0d-4227-9195-7daa9d044919-config\") pod \"route-controller-manager-8b9b5687d-wkb5r\" (UID: \"92cf2506-fc0d-4227-9195-7daa9d044919\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.758424 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9xj9\" (UniqueName: \"kubernetes.io/projected/92cf2506-fc0d-4227-9195-7daa9d044919-kube-api-access-x9xj9\") pod \"route-controller-manager-8b9b5687d-wkb5r\" (UID: \"92cf2506-fc0d-4227-9195-7daa9d044919\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.859882 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9xj9\" (UniqueName: \"kubernetes.io/projected/92cf2506-fc0d-4227-9195-7daa9d044919-kube-api-access-x9xj9\") pod \"route-controller-manager-8b9b5687d-wkb5r\" (UID: \"92cf2506-fc0d-4227-9195-7daa9d044919\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.860354 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92cf2506-fc0d-4227-9195-7daa9d044919-serving-cert\") pod \"route-controller-manager-8b9b5687d-wkb5r\" (UID: \"92cf2506-fc0d-4227-9195-7daa9d044919\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.860615 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92cf2506-fc0d-4227-9195-7daa9d044919-client-ca\") pod \"route-controller-manager-8b9b5687d-wkb5r\" (UID: \"92cf2506-fc0d-4227-9195-7daa9d044919\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.860895 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92cf2506-fc0d-4227-9195-7daa9d044919-config\") pod \"route-controller-manager-8b9b5687d-wkb5r\" (UID: \"92cf2506-fc0d-4227-9195-7daa9d044919\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.862514 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92cf2506-fc0d-4227-9195-7daa9d044919-client-ca\") pod \"route-controller-manager-8b9b5687d-wkb5r\" (UID: \"92cf2506-fc0d-4227-9195-7daa9d044919\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.863300 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92cf2506-fc0d-4227-9195-7daa9d044919-config\") pod \"route-controller-manager-8b9b5687d-wkb5r\" (UID: \"92cf2506-fc0d-4227-9195-7daa9d044919\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.877530 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92cf2506-fc0d-4227-9195-7daa9d044919-serving-cert\") pod \"route-controller-manager-8b9b5687d-wkb5r\" (UID: \"92cf2506-fc0d-4227-9195-7daa9d044919\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.893788 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9xj9\" (UniqueName: \"kubernetes.io/projected/92cf2506-fc0d-4227-9195-7daa9d044919-kube-api-access-x9xj9\") pod \"route-controller-manager-8b9b5687d-wkb5r\" (UID: \"92cf2506-fc0d-4227-9195-7daa9d044919\") " pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" Feb 27 16:30:37 crc kubenswrapper[4814]: I0227 16:30:37.993648 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" Feb 27 16:30:38 crc kubenswrapper[4814]: I0227 16:30:38.495096 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42bd9983-08bf-4263-8fcb-0714d5067d3a" path="/var/lib/kubelet/pods/42bd9983-08bf-4263-8fcb-0714d5067d3a/volumes" Feb 27 16:30:39 crc kubenswrapper[4814]: I0227 16:30:39.132855 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r"] Feb 27 16:30:39 crc kubenswrapper[4814]: W0227 16:30:39.149179 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92cf2506_fc0d_4227_9195_7daa9d044919.slice/crio-780eb778467f9f8db75c63aadbf51e8705943a6461362542295312e8a10efaec WatchSource:0}: Error finding container 780eb778467f9f8db75c63aadbf51e8705943a6461362542295312e8a10efaec: Status 404 returned error can't find the container with id 780eb778467f9f8db75c63aadbf51e8705943a6461362542295312e8a10efaec Feb 27 16:30:39 crc kubenswrapper[4814]: I0227 16:30:39.577884 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" event={"ID":"92cf2506-fc0d-4227-9195-7daa9d044919","Type":"ContainerStarted","Data":"cc2910e4b8adb3a01d3153f4eb638d0f4b8c0961caf5a8a4a10655b488a4b7d7"} Feb 27 16:30:39 crc kubenswrapper[4814]: I0227 16:30:39.578543 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" Feb 27 16:30:39 crc kubenswrapper[4814]: I0227 16:30:39.578566 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" event={"ID":"92cf2506-fc0d-4227-9195-7daa9d044919","Type":"ContainerStarted","Data":"780eb778467f9f8db75c63aadbf51e8705943a6461362542295312e8a10efaec"} Feb 27 16:30:39 crc kubenswrapper[4814]: I0227 16:30:39.623585 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" podStartSLOduration=4.623532073 podStartE2EDuration="4.623532073s" podCreationTimestamp="2026-02-27 16:30:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:30:39.599145951 +0000 UTC m=+452.051770821" watchObservedRunningTime="2026-02-27 16:30:39.623532073 +0000 UTC m=+452.076156903" Feb 27 16:30:39 crc kubenswrapper[4814]: I0227 16:30:39.886090 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8b9b5687d-wkb5r" Feb 27 16:30:43 crc kubenswrapper[4814]: I0227 16:30:43.906589 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cm8p5"] Feb 27 16:30:43 crc kubenswrapper[4814]: I0227 16:30:43.907687 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cm8p5" podUID="f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" containerName="registry-server" containerID="cri-o://87931504b222c833ffc5fd5cb8dc55a426269355d7d33123a4f5c0c883e15d3a" gracePeriod=30 Feb 27 16:30:43 crc kubenswrapper[4814]: I0227 16:30:43.928938 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sk6rs"] Feb 27 16:30:43 crc kubenswrapper[4814]: I0227 16:30:43.929487 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sk6rs" podUID="41b851a0-9bda-4e20-a692-5f796f55a749" containerName="registry-server" containerID="cri-o://368c23bbb8809ac49a66f06ee8351d5dab07d09c3f389c98854f5e3a6b21c5dd" gracePeriod=30 Feb 27 16:30:43 crc kubenswrapper[4814]: I0227 16:30:43.934770 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mjlld"] Feb 27 16:30:43 crc kubenswrapper[4814]: I0227 16:30:43.935242 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" podUID="dfa49cf2-dc52-4c27-8486-ec1efec61744" containerName="marketplace-operator" containerID="cri-o://fe52ca82eeb498fd270e34ebb923c23f179ca8796e08c87c33acf686f95502db" gracePeriod=30 Feb 27 16:30:43 crc kubenswrapper[4814]: I0227 16:30:43.938704 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4zbl6"] Feb 27 16:30:43 crc kubenswrapper[4814]: I0227 16:30:43.939018 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4zbl6" podUID="5df9c141-ab4f-40e3-b733-da2a2712a326" containerName="registry-server" containerID="cri-o://81e8fe574c70824a561cd7739c591fe5f5ec131c54b59a261eb90b69a3f973d0" gracePeriod=30 Feb 27 16:30:43 crc kubenswrapper[4814]: I0227 16:30:43.955970 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lvrr5"] Feb 27 16:30:43 crc kubenswrapper[4814]: I0227 16:30:43.957082 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-lvrr5" Feb 27 16:30:43 crc kubenswrapper[4814]: I0227 16:30:43.969379 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9ftgk"] Feb 27 16:30:43 crc kubenswrapper[4814]: I0227 16:30:43.969720 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9ftgk" podUID="b80cc2cd-d07d-4225-b078-9c70d8e189c7" containerName="registry-server" containerID="cri-o://42074e214640646aa655a689cfe857d210987bcec894e98a7444ec78886ef8ec" gracePeriod=30 Feb 27 16:30:43 crc kubenswrapper[4814]: I0227 16:30:43.972756 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lvrr5"] Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.070099 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5d324fb-1565-4d40-9304-9e2f6adc9604-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-lvrr5\" (UID: \"c5d324fb-1565-4d40-9304-9e2f6adc9604\") " pod="openshift-marketplace/marketplace-operator-79b997595-lvrr5" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.070203 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c5d324fb-1565-4d40-9304-9e2f6adc9604-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-lvrr5\" (UID: \"c5d324fb-1565-4d40-9304-9e2f6adc9604\") " pod="openshift-marketplace/marketplace-operator-79b997595-lvrr5" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.070272 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqvx4\" (UniqueName: \"kubernetes.io/projected/c5d324fb-1565-4d40-9304-9e2f6adc9604-kube-api-access-kqvx4\") pod \"marketplace-operator-79b997595-lvrr5\" (UID: \"c5d324fb-1565-4d40-9304-9e2f6adc9604\") " pod="openshift-marketplace/marketplace-operator-79b997595-lvrr5" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.172373 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c5d324fb-1565-4d40-9304-9e2f6adc9604-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-lvrr5\" (UID: \"c5d324fb-1565-4d40-9304-9e2f6adc9604\") " pod="openshift-marketplace/marketplace-operator-79b997595-lvrr5" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.172896 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqvx4\" (UniqueName: \"kubernetes.io/projected/c5d324fb-1565-4d40-9304-9e2f6adc9604-kube-api-access-kqvx4\") pod \"marketplace-operator-79b997595-lvrr5\" (UID: \"c5d324fb-1565-4d40-9304-9e2f6adc9604\") " pod="openshift-marketplace/marketplace-operator-79b997595-lvrr5" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.172945 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5d324fb-1565-4d40-9304-9e2f6adc9604-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-lvrr5\" (UID: \"c5d324fb-1565-4d40-9304-9e2f6adc9604\") " pod="openshift-marketplace/marketplace-operator-79b997595-lvrr5" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.174161 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5d324fb-1565-4d40-9304-9e2f6adc9604-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-lvrr5\" (UID: \"c5d324fb-1565-4d40-9304-9e2f6adc9604\") " pod="openshift-marketplace/marketplace-operator-79b997595-lvrr5" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.190730 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqvx4\" (UniqueName: \"kubernetes.io/projected/c5d324fb-1565-4d40-9304-9e2f6adc9604-kube-api-access-kqvx4\") pod \"marketplace-operator-79b997595-lvrr5\" (UID: \"c5d324fb-1565-4d40-9304-9e2f6adc9604\") " pod="openshift-marketplace/marketplace-operator-79b997595-lvrr5" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.195527 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c5d324fb-1565-4d40-9304-9e2f6adc9604-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-lvrr5\" (UID: \"c5d324fb-1565-4d40-9304-9e2f6adc9604\") " pod="openshift-marketplace/marketplace-operator-79b997595-lvrr5" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.287155 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-lvrr5" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.558935 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.640989 4814 generic.go:334] "Generic (PLEG): container finished" podID="5df9c141-ab4f-40e3-b733-da2a2712a326" containerID="81e8fe574c70824a561cd7739c591fe5f5ec131c54b59a261eb90b69a3f973d0" exitCode=0 Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.641083 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zbl6" event={"ID":"5df9c141-ab4f-40e3-b733-da2a2712a326","Type":"ContainerDied","Data":"81e8fe574c70824a561cd7739c591fe5f5ec131c54b59a261eb90b69a3f973d0"} Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.643123 4814 generic.go:334] "Generic (PLEG): container finished" podID="f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" containerID="87931504b222c833ffc5fd5cb8dc55a426269355d7d33123a4f5c0c883e15d3a" exitCode=0 Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.643194 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cm8p5" event={"ID":"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375","Type":"ContainerDied","Data":"87931504b222c833ffc5fd5cb8dc55a426269355d7d33123a4f5c0c883e15d3a"} Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.643219 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cm8p5" event={"ID":"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375","Type":"ContainerDied","Data":"6683dda6eb1039822dea692ee60f4fd262751c4125c7f2bad504712dcece7f7d"} Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.643271 4814 scope.go:117] "RemoveContainer" containerID="87931504b222c833ffc5fd5cb8dc55a426269355d7d33123a4f5c0c883e15d3a" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.643497 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cm8p5" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.653814 4814 generic.go:334] "Generic (PLEG): container finished" podID="b80cc2cd-d07d-4225-b078-9c70d8e189c7" containerID="42074e214640646aa655a689cfe857d210987bcec894e98a7444ec78886ef8ec" exitCode=0 Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.653877 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ftgk" event={"ID":"b80cc2cd-d07d-4225-b078-9c70d8e189c7","Type":"ContainerDied","Data":"42074e214640646aa655a689cfe857d210987bcec894e98a7444ec78886ef8ec"} Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.659593 4814 generic.go:334] "Generic (PLEG): container finished" podID="41b851a0-9bda-4e20-a692-5f796f55a749" containerID="368c23bbb8809ac49a66f06ee8351d5dab07d09c3f389c98854f5e3a6b21c5dd" exitCode=0 Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.659684 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sk6rs" event={"ID":"41b851a0-9bda-4e20-a692-5f796f55a749","Type":"ContainerDied","Data":"368c23bbb8809ac49a66f06ee8351d5dab07d09c3f389c98854f5e3a6b21c5dd"} Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.663687 4814 generic.go:334] "Generic (PLEG): container finished" podID="dfa49cf2-dc52-4c27-8486-ec1efec61744" containerID="fe52ca82eeb498fd270e34ebb923c23f179ca8796e08c87c33acf686f95502db" exitCode=0 Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.663854 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" event={"ID":"dfa49cf2-dc52-4c27-8486-ec1efec61744","Type":"ContainerDied","Data":"fe52ca82eeb498fd270e34ebb923c23f179ca8796e08c87c33acf686f95502db"} Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.676198 4814 scope.go:117] "RemoveContainer" containerID="208fbbd20e9ec61739404ddb25a8913c15f9d97180bddccb121fd850e60545c0" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.686650 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-catalog-content\") pod \"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375\" (UID: \"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375\") " Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.686698 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-utilities\") pod \"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375\" (UID: \"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375\") " Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.686805 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5cld\" (UniqueName: \"kubernetes.io/projected/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-kube-api-access-h5cld\") pod \"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375\" (UID: \"f1ed29a3-a2f5-46e1-ac86-b3adc8c52375\") " Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.688136 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-utilities" (OuterVolumeSpecName: "utilities") pod "f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" (UID: "f1ed29a3-a2f5-46e1-ac86-b3adc8c52375"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.695696 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-kube-api-access-h5cld" (OuterVolumeSpecName: "kube-api-access-h5cld") pod "f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" (UID: "f1ed29a3-a2f5-46e1-ac86-b3adc8c52375"). InnerVolumeSpecName "kube-api-access-h5cld". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.704973 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.712697 4814 scope.go:117] "RemoveContainer" containerID="c56f2147f422717aab783873ae75de9d5cdc15359a4cd3f56720360d84a63586" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.738513 4814 scope.go:117] "RemoveContainer" containerID="87931504b222c833ffc5fd5cb8dc55a426269355d7d33123a4f5c0c883e15d3a" Feb 27 16:30:44 crc kubenswrapper[4814]: E0227 16:30:44.739064 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87931504b222c833ffc5fd5cb8dc55a426269355d7d33123a4f5c0c883e15d3a\": container with ID starting with 87931504b222c833ffc5fd5cb8dc55a426269355d7d33123a4f5c0c883e15d3a not found: ID does not exist" containerID="87931504b222c833ffc5fd5cb8dc55a426269355d7d33123a4f5c0c883e15d3a" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.739118 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87931504b222c833ffc5fd5cb8dc55a426269355d7d33123a4f5c0c883e15d3a"} err="failed to get container status \"87931504b222c833ffc5fd5cb8dc55a426269355d7d33123a4f5c0c883e15d3a\": rpc error: code = NotFound desc = could not find container \"87931504b222c833ffc5fd5cb8dc55a426269355d7d33123a4f5c0c883e15d3a\": container with ID starting with 87931504b222c833ffc5fd5cb8dc55a426269355d7d33123a4f5c0c883e15d3a not found: ID does not exist" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.739149 4814 scope.go:117] "RemoveContainer" containerID="208fbbd20e9ec61739404ddb25a8913c15f9d97180bddccb121fd850e60545c0" Feb 27 16:30:44 crc kubenswrapper[4814]: E0227 16:30:44.739473 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"208fbbd20e9ec61739404ddb25a8913c15f9d97180bddccb121fd850e60545c0\": container with ID starting with 208fbbd20e9ec61739404ddb25a8913c15f9d97180bddccb121fd850e60545c0 not found: ID does not exist" containerID="208fbbd20e9ec61739404ddb25a8913c15f9d97180bddccb121fd850e60545c0" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.739506 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"208fbbd20e9ec61739404ddb25a8913c15f9d97180bddccb121fd850e60545c0"} err="failed to get container status \"208fbbd20e9ec61739404ddb25a8913c15f9d97180bddccb121fd850e60545c0\": rpc error: code = NotFound desc = could not find container \"208fbbd20e9ec61739404ddb25a8913c15f9d97180bddccb121fd850e60545c0\": container with ID starting with 208fbbd20e9ec61739404ddb25a8913c15f9d97180bddccb121fd850e60545c0 not found: ID does not exist" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.739533 4814 scope.go:117] "RemoveContainer" containerID="c56f2147f422717aab783873ae75de9d5cdc15359a4cd3f56720360d84a63586" Feb 27 16:30:44 crc kubenswrapper[4814]: E0227 16:30:44.739834 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c56f2147f422717aab783873ae75de9d5cdc15359a4cd3f56720360d84a63586\": container with ID starting with c56f2147f422717aab783873ae75de9d5cdc15359a4cd3f56720360d84a63586 not found: ID does not exist" containerID="c56f2147f422717aab783873ae75de9d5cdc15359a4cd3f56720360d84a63586" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.739871 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c56f2147f422717aab783873ae75de9d5cdc15359a4cd3f56720360d84a63586"} err="failed to get container status \"c56f2147f422717aab783873ae75de9d5cdc15359a4cd3f56720360d84a63586\": rpc error: code = NotFound desc = could not find container \"c56f2147f422717aab783873ae75de9d5cdc15359a4cd3f56720360d84a63586\": container with ID starting with c56f2147f422717aab783873ae75de9d5cdc15359a4cd3f56720360d84a63586 not found: ID does not exist" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.739899 4814 scope.go:117] "RemoveContainer" containerID="c368bbd9a1ae5371c62ca4d7ecc6b992fe4809aa77ddb9bd665461038649f222" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.750685 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.757107 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" (UID: "f1ed29a3-a2f5-46e1-ac86-b3adc8c52375"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.787830 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmsmw\" (UniqueName: \"kubernetes.io/projected/dfa49cf2-dc52-4c27-8486-ec1efec61744-kube-api-access-qmsmw\") pod \"dfa49cf2-dc52-4c27-8486-ec1efec61744\" (UID: \"dfa49cf2-dc52-4c27-8486-ec1efec61744\") " Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.787903 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dfa49cf2-dc52-4c27-8486-ec1efec61744-marketplace-operator-metrics\") pod \"dfa49cf2-dc52-4c27-8486-ec1efec61744\" (UID: \"dfa49cf2-dc52-4c27-8486-ec1efec61744\") " Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.788048 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfa49cf2-dc52-4c27-8486-ec1efec61744-marketplace-trusted-ca\") pod \"dfa49cf2-dc52-4c27-8486-ec1efec61744\" (UID: \"dfa49cf2-dc52-4c27-8486-ec1efec61744\") " Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.788887 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfa49cf2-dc52-4c27-8486-ec1efec61744-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "dfa49cf2-dc52-4c27-8486-ec1efec61744" (UID: "dfa49cf2-dc52-4c27-8486-ec1efec61744"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.789246 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5cld\" (UniqueName: \"kubernetes.io/projected/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-kube-api-access-h5cld\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.789327 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.789338 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.789349 4814 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfa49cf2-dc52-4c27-8486-ec1efec61744-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.795306 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfa49cf2-dc52-4c27-8486-ec1efec61744-kube-api-access-qmsmw" (OuterVolumeSpecName: "kube-api-access-qmsmw") pod "dfa49cf2-dc52-4c27-8486-ec1efec61744" (UID: "dfa49cf2-dc52-4c27-8486-ec1efec61744"). InnerVolumeSpecName "kube-api-access-qmsmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.795528 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfa49cf2-dc52-4c27-8486-ec1efec61744-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "dfa49cf2-dc52-4c27-8486-ec1efec61744" (UID: "dfa49cf2-dc52-4c27-8486-ec1efec61744"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.796993 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.813199 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.890748 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5df9c141-ab4f-40e3-b733-da2a2712a326-catalog-content\") pod \"5df9c141-ab4f-40e3-b733-da2a2712a326\" (UID: \"5df9c141-ab4f-40e3-b733-da2a2712a326\") " Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.890781 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kts7\" (UniqueName: \"kubernetes.io/projected/b80cc2cd-d07d-4225-b078-9c70d8e189c7-kube-api-access-9kts7\") pod \"b80cc2cd-d07d-4225-b078-9c70d8e189c7\" (UID: \"b80cc2cd-d07d-4225-b078-9c70d8e189c7\") " Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.891329 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpb22\" (UniqueName: \"kubernetes.io/projected/5df9c141-ab4f-40e3-b733-da2a2712a326-kube-api-access-tpb22\") pod \"5df9c141-ab4f-40e3-b733-da2a2712a326\" (UID: \"5df9c141-ab4f-40e3-b733-da2a2712a326\") " Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.891355 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5df9c141-ab4f-40e3-b733-da2a2712a326-utilities\") pod \"5df9c141-ab4f-40e3-b733-da2a2712a326\" (UID: \"5df9c141-ab4f-40e3-b733-da2a2712a326\") " Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.891394 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b80cc2cd-d07d-4225-b078-9c70d8e189c7-catalog-content\") pod \"b80cc2cd-d07d-4225-b078-9c70d8e189c7\" (UID: \"b80cc2cd-d07d-4225-b078-9c70d8e189c7\") " Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.891416 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41b851a0-9bda-4e20-a692-5f796f55a749-utilities\") pod \"41b851a0-9bda-4e20-a692-5f796f55a749\" (UID: \"41b851a0-9bda-4e20-a692-5f796f55a749\") " Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.891438 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41b851a0-9bda-4e20-a692-5f796f55a749-catalog-content\") pod \"41b851a0-9bda-4e20-a692-5f796f55a749\" (UID: \"41b851a0-9bda-4e20-a692-5f796f55a749\") " Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.891468 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b80cc2cd-d07d-4225-b078-9c70d8e189c7-utilities\") pod \"b80cc2cd-d07d-4225-b078-9c70d8e189c7\" (UID: \"b80cc2cd-d07d-4225-b078-9c70d8e189c7\") " Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.891484 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgd64\" (UniqueName: \"kubernetes.io/projected/41b851a0-9bda-4e20-a692-5f796f55a749-kube-api-access-vgd64\") pod \"41b851a0-9bda-4e20-a692-5f796f55a749\" (UID: \"41b851a0-9bda-4e20-a692-5f796f55a749\") " Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.891967 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmsmw\" (UniqueName: \"kubernetes.io/projected/dfa49cf2-dc52-4c27-8486-ec1efec61744-kube-api-access-qmsmw\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.891985 4814 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dfa49cf2-dc52-4c27-8486-ec1efec61744-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.892938 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b80cc2cd-d07d-4225-b078-9c70d8e189c7-utilities" (OuterVolumeSpecName: "utilities") pod "b80cc2cd-d07d-4225-b078-9c70d8e189c7" (UID: "b80cc2cd-d07d-4225-b078-9c70d8e189c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.895212 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b80cc2cd-d07d-4225-b078-9c70d8e189c7-kube-api-access-9kts7" (OuterVolumeSpecName: "kube-api-access-9kts7") pod "b80cc2cd-d07d-4225-b078-9c70d8e189c7" (UID: "b80cc2cd-d07d-4225-b078-9c70d8e189c7"). InnerVolumeSpecName "kube-api-access-9kts7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.896454 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41b851a0-9bda-4e20-a692-5f796f55a749-kube-api-access-vgd64" (OuterVolumeSpecName: "kube-api-access-vgd64") pod "41b851a0-9bda-4e20-a692-5f796f55a749" (UID: "41b851a0-9bda-4e20-a692-5f796f55a749"). InnerVolumeSpecName "kube-api-access-vgd64". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.900116 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5df9c141-ab4f-40e3-b733-da2a2712a326-kube-api-access-tpb22" (OuterVolumeSpecName: "kube-api-access-tpb22") pod "5df9c141-ab4f-40e3-b733-da2a2712a326" (UID: "5df9c141-ab4f-40e3-b733-da2a2712a326"). InnerVolumeSpecName "kube-api-access-tpb22". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.908210 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41b851a0-9bda-4e20-a692-5f796f55a749-utilities" (OuterVolumeSpecName: "utilities") pod "41b851a0-9bda-4e20-a692-5f796f55a749" (UID: "41b851a0-9bda-4e20-a692-5f796f55a749"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.915898 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5df9c141-ab4f-40e3-b733-da2a2712a326-utilities" (OuterVolumeSpecName: "utilities") pod "5df9c141-ab4f-40e3-b733-da2a2712a326" (UID: "5df9c141-ab4f-40e3-b733-da2a2712a326"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.918543 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5df9c141-ab4f-40e3-b733-da2a2712a326-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5df9c141-ab4f-40e3-b733-da2a2712a326" (UID: "5df9c141-ab4f-40e3-b733-da2a2712a326"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.925861 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lvrr5"] Feb 27 16:30:44 crc kubenswrapper[4814]: W0227 16:30:44.928191 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5d324fb_1565_4d40_9304_9e2f6adc9604.slice/crio-758a189e3189205e8981c1ac9216ff04692cc3c38687b0d28a19d4095328f753 WatchSource:0}: Error finding container 758a189e3189205e8981c1ac9216ff04692cc3c38687b0d28a19d4095328f753: Status 404 returned error can't find the container with id 758a189e3189205e8981c1ac9216ff04692cc3c38687b0d28a19d4095328f753 Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.989726 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41b851a0-9bda-4e20-a692-5f796f55a749-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "41b851a0-9bda-4e20-a692-5f796f55a749" (UID: "41b851a0-9bda-4e20-a692-5f796f55a749"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.992883 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41b851a0-9bda-4e20-a692-5f796f55a749-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.992912 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41b851a0-9bda-4e20-a692-5f796f55a749-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.992928 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b80cc2cd-d07d-4225-b078-9c70d8e189c7-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.992958 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgd64\" (UniqueName: \"kubernetes.io/projected/41b851a0-9bda-4e20-a692-5f796f55a749-kube-api-access-vgd64\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.992970 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5df9c141-ab4f-40e3-b733-da2a2712a326-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.992981 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kts7\" (UniqueName: \"kubernetes.io/projected/b80cc2cd-d07d-4225-b078-9c70d8e189c7-kube-api-access-9kts7\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.992993 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpb22\" (UniqueName: \"kubernetes.io/projected/5df9c141-ab4f-40e3-b733-da2a2712a326-kube-api-access-tpb22\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.993003 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5df9c141-ab4f-40e3-b733-da2a2712a326-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:44 crc kubenswrapper[4814]: I0227 16:30:44.996215 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cm8p5"] Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:44.999869 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cm8p5"] Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.051521 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b80cc2cd-d07d-4225-b078-9c70d8e189c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b80cc2cd-d07d-4225-b078-9c70d8e189c7" (UID: "b80cc2cd-d07d-4225-b078-9c70d8e189c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.093884 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b80cc2cd-d07d-4225-b078-9c70d8e189c7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.673218 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ftgk" event={"ID":"b80cc2cd-d07d-4225-b078-9c70d8e189c7","Type":"ContainerDied","Data":"498c9c32b480942bf3b30764e5cab7578b22e01bc4773bc209332f9e453d4b80"} Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.673383 4814 scope.go:117] "RemoveContainer" containerID="42074e214640646aa655a689cfe857d210987bcec894e98a7444ec78886ef8ec" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.673466 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ftgk" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.675853 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-lvrr5" event={"ID":"c5d324fb-1565-4d40-9304-9e2f6adc9604","Type":"ContainerStarted","Data":"65d370368851c14c7e8c7cdb934c000b321a7629b3ecbfbe043da38cf61e6a25"} Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.675965 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-lvrr5" event={"ID":"c5d324fb-1565-4d40-9304-9e2f6adc9604","Type":"ContainerStarted","Data":"758a189e3189205e8981c1ac9216ff04692cc3c38687b0d28a19d4095328f753"} Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.676046 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-lvrr5" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.680102 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-lvrr5" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.683799 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sk6rs" event={"ID":"41b851a0-9bda-4e20-a692-5f796f55a749","Type":"ContainerDied","Data":"4f549b4cfbcf24cc2938fb684e94cf90a0493916225b398319d177632e456eab"} Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.683888 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sk6rs" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.690653 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" event={"ID":"dfa49cf2-dc52-4c27-8486-ec1efec61744","Type":"ContainerDied","Data":"436c02e76e48674afb5902acd48631763f9a4ab4ccd2d5dc2c880d50a8814641"} Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.691212 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mjlld" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.695073 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zbl6" event={"ID":"5df9c141-ab4f-40e3-b733-da2a2712a326","Type":"ContainerDied","Data":"85ec660e523122f9e380a271b94600a98372d92e39c7c0ccca9cf657f8a08811"} Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.695250 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4zbl6" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.711624 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-lvrr5" podStartSLOduration=2.711592924 podStartE2EDuration="2.711592924s" podCreationTimestamp="2026-02-27 16:30:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:30:45.70530065 +0000 UTC m=+458.157925480" watchObservedRunningTime="2026-02-27 16:30:45.711592924 +0000 UTC m=+458.164217794" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.712733 4814 scope.go:117] "RemoveContainer" containerID="c93f06679f23c4e6585f42ea61bdf308945fe55fac7aeb6a0b8f3afe9c640c8f" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.771121 4814 scope.go:117] "RemoveContainer" containerID="b5b291085889d8daaec79af7b1fcb2f89474bd83feeac9c571d9d17f26147820" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.787345 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sk6rs"] Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.792682 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sk6rs"] Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.796625 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9ftgk"] Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.798466 4814 scope.go:117] "RemoveContainer" containerID="368c23bbb8809ac49a66f06ee8351d5dab07d09c3f389c98854f5e3a6b21c5dd" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.802871 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9ftgk"] Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.808467 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mjlld"] Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.812408 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mjlld"] Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.820341 4814 scope.go:117] "RemoveContainer" containerID="07575168eaf9cbb5d7cc834dc2a531a5cb850e0efe81d9f5fffeb1286239e917" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.824546 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4zbl6"] Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.828919 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4zbl6"] Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.837793 4814 scope.go:117] "RemoveContainer" containerID="c91bbb45267bcaa1b07196607bd447558099939728a7922c29ef04c8ec6303b7" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.850827 4814 scope.go:117] "RemoveContainer" containerID="fe52ca82eeb498fd270e34ebb923c23f179ca8796e08c87c33acf686f95502db" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.869857 4814 scope.go:117] "RemoveContainer" containerID="81e8fe574c70824a561cd7739c591fe5f5ec131c54b59a261eb90b69a3f973d0" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.882124 4814 scope.go:117] "RemoveContainer" containerID="74ba1ee457e00edd1799583460cb1f1d4bae89ff1fb268ce58bff562a9ff515f" Feb 27 16:30:45 crc kubenswrapper[4814]: I0227 16:30:45.898748 4814 scope.go:117] "RemoveContainer" containerID="adf15ca9cd416ed78e0bd063bb50774a92c7dd0047a7cbd501b6805bf31e9f46" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.315417 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r5rrh"] Feb 27 16:30:46 crc kubenswrapper[4814]: E0227 16:30:46.315819 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" containerName="extract-utilities" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.315844 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" containerName="extract-utilities" Feb 27 16:30:46 crc kubenswrapper[4814]: E0227 16:30:46.315863 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" containerName="extract-content" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.315875 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" containerName="extract-content" Feb 27 16:30:46 crc kubenswrapper[4814]: E0227 16:30:46.315893 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5df9c141-ab4f-40e3-b733-da2a2712a326" containerName="registry-server" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.315906 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="5df9c141-ab4f-40e3-b733-da2a2712a326" containerName="registry-server" Feb 27 16:30:46 crc kubenswrapper[4814]: E0227 16:30:46.315918 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5df9c141-ab4f-40e3-b733-da2a2712a326" containerName="extract-utilities" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.315929 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="5df9c141-ab4f-40e3-b733-da2a2712a326" containerName="extract-utilities" Feb 27 16:30:46 crc kubenswrapper[4814]: E0227 16:30:46.315945 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80cc2cd-d07d-4225-b078-9c70d8e189c7" containerName="extract-utilities" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.315957 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80cc2cd-d07d-4225-b078-9c70d8e189c7" containerName="extract-utilities" Feb 27 16:30:46 crc kubenswrapper[4814]: E0227 16:30:46.315970 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b851a0-9bda-4e20-a692-5f796f55a749" containerName="registry-server" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.315981 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b851a0-9bda-4e20-a692-5f796f55a749" containerName="registry-server" Feb 27 16:30:46 crc kubenswrapper[4814]: E0227 16:30:46.316002 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5df9c141-ab4f-40e3-b733-da2a2712a326" containerName="extract-content" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.316013 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="5df9c141-ab4f-40e3-b733-da2a2712a326" containerName="extract-content" Feb 27 16:30:46 crc kubenswrapper[4814]: E0227 16:30:46.316025 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" containerName="registry-server" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.316035 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" containerName="registry-server" Feb 27 16:30:46 crc kubenswrapper[4814]: E0227 16:30:46.316049 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfa49cf2-dc52-4c27-8486-ec1efec61744" containerName="marketplace-operator" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.316060 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfa49cf2-dc52-4c27-8486-ec1efec61744" containerName="marketplace-operator" Feb 27 16:30:46 crc kubenswrapper[4814]: E0227 16:30:46.316072 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80cc2cd-d07d-4225-b078-9c70d8e189c7" containerName="registry-server" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.316083 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80cc2cd-d07d-4225-b078-9c70d8e189c7" containerName="registry-server" Feb 27 16:30:46 crc kubenswrapper[4814]: E0227 16:30:46.316098 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b851a0-9bda-4e20-a692-5f796f55a749" containerName="extract-utilities" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.316108 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b851a0-9bda-4e20-a692-5f796f55a749" containerName="extract-utilities" Feb 27 16:30:46 crc kubenswrapper[4814]: E0227 16:30:46.316126 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80cc2cd-d07d-4225-b078-9c70d8e189c7" containerName="extract-content" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.316137 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80cc2cd-d07d-4225-b078-9c70d8e189c7" containerName="extract-content" Feb 27 16:30:46 crc kubenswrapper[4814]: E0227 16:30:46.316151 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b851a0-9bda-4e20-a692-5f796f55a749" containerName="extract-content" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.316161 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b851a0-9bda-4e20-a692-5f796f55a749" containerName="extract-content" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.316326 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfa49cf2-dc52-4c27-8486-ec1efec61744" containerName="marketplace-operator" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.316345 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="5df9c141-ab4f-40e3-b733-da2a2712a326" containerName="registry-server" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.316358 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="b80cc2cd-d07d-4225-b078-9c70d8e189c7" containerName="registry-server" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.316372 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" containerName="registry-server" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.316384 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="41b851a0-9bda-4e20-a692-5f796f55a749" containerName="registry-server" Feb 27 16:30:46 crc kubenswrapper[4814]: E0227 16:30:46.316534 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfa49cf2-dc52-4c27-8486-ec1efec61744" containerName="marketplace-operator" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.316550 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfa49cf2-dc52-4c27-8486-ec1efec61744" containerName="marketplace-operator" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.316703 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfa49cf2-dc52-4c27-8486-ec1efec61744" containerName="marketplace-operator" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.318043 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5rrh" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.324560 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.337834 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r5rrh"] Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.416552 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5468\" (UniqueName: \"kubernetes.io/projected/73643f07-8d1a-4321-902c-12a52a2fe4da-kube-api-access-f5468\") pod \"redhat-operators-r5rrh\" (UID: \"73643f07-8d1a-4321-902c-12a52a2fe4da\") " pod="openshift-marketplace/redhat-operators-r5rrh" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.416663 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73643f07-8d1a-4321-902c-12a52a2fe4da-catalog-content\") pod \"redhat-operators-r5rrh\" (UID: \"73643f07-8d1a-4321-902c-12a52a2fe4da\") " pod="openshift-marketplace/redhat-operators-r5rrh" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.416701 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73643f07-8d1a-4321-902c-12a52a2fe4da-utilities\") pod \"redhat-operators-r5rrh\" (UID: \"73643f07-8d1a-4321-902c-12a52a2fe4da\") " pod="openshift-marketplace/redhat-operators-r5rrh" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.495504 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41b851a0-9bda-4e20-a692-5f796f55a749" path="/var/lib/kubelet/pods/41b851a0-9bda-4e20-a692-5f796f55a749/volumes" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.496124 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5df9c141-ab4f-40e3-b733-da2a2712a326" path="/var/lib/kubelet/pods/5df9c141-ab4f-40e3-b733-da2a2712a326/volumes" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.496729 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b80cc2cd-d07d-4225-b078-9c70d8e189c7" path="/var/lib/kubelet/pods/b80cc2cd-d07d-4225-b078-9c70d8e189c7/volumes" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.497785 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfa49cf2-dc52-4c27-8486-ec1efec61744" path="/var/lib/kubelet/pods/dfa49cf2-dc52-4c27-8486-ec1efec61744/volumes" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.498298 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1ed29a3-a2f5-46e1-ac86-b3adc8c52375" path="/var/lib/kubelet/pods/f1ed29a3-a2f5-46e1-ac86-b3adc8c52375/volumes" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.518128 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5468\" (UniqueName: \"kubernetes.io/projected/73643f07-8d1a-4321-902c-12a52a2fe4da-kube-api-access-f5468\") pod \"redhat-operators-r5rrh\" (UID: \"73643f07-8d1a-4321-902c-12a52a2fe4da\") " pod="openshift-marketplace/redhat-operators-r5rrh" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.518219 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73643f07-8d1a-4321-902c-12a52a2fe4da-catalog-content\") pod \"redhat-operators-r5rrh\" (UID: \"73643f07-8d1a-4321-902c-12a52a2fe4da\") " pod="openshift-marketplace/redhat-operators-r5rrh" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.518296 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73643f07-8d1a-4321-902c-12a52a2fe4da-utilities\") pod \"redhat-operators-r5rrh\" (UID: \"73643f07-8d1a-4321-902c-12a52a2fe4da\") " pod="openshift-marketplace/redhat-operators-r5rrh" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.518746 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73643f07-8d1a-4321-902c-12a52a2fe4da-utilities\") pod \"redhat-operators-r5rrh\" (UID: \"73643f07-8d1a-4321-902c-12a52a2fe4da\") " pod="openshift-marketplace/redhat-operators-r5rrh" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.518805 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73643f07-8d1a-4321-902c-12a52a2fe4da-catalog-content\") pod \"redhat-operators-r5rrh\" (UID: \"73643f07-8d1a-4321-902c-12a52a2fe4da\") " pod="openshift-marketplace/redhat-operators-r5rrh" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.544223 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5468\" (UniqueName: \"kubernetes.io/projected/73643f07-8d1a-4321-902c-12a52a2fe4da-kube-api-access-f5468\") pod \"redhat-operators-r5rrh\" (UID: \"73643f07-8d1a-4321-902c-12a52a2fe4da\") " pod="openshift-marketplace/redhat-operators-r5rrh" Feb 27 16:30:46 crc kubenswrapper[4814]: I0227 16:30:46.659287 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5rrh" Feb 27 16:30:47 crc kubenswrapper[4814]: I0227 16:30:47.083859 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r5rrh"] Feb 27 16:30:47 crc kubenswrapper[4814]: I0227 16:30:47.736622 4814 generic.go:334] "Generic (PLEG): container finished" podID="73643f07-8d1a-4321-902c-12a52a2fe4da" containerID="61d87bb0bd75cc00f7b0c8880fb3a85b7af67113e48f13e04fd76562926088ff" exitCode=0 Feb 27 16:30:47 crc kubenswrapper[4814]: I0227 16:30:47.736703 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5rrh" event={"ID":"73643f07-8d1a-4321-902c-12a52a2fe4da","Type":"ContainerDied","Data":"61d87bb0bd75cc00f7b0c8880fb3a85b7af67113e48f13e04fd76562926088ff"} Feb 27 16:30:47 crc kubenswrapper[4814]: I0227 16:30:47.736766 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5rrh" event={"ID":"73643f07-8d1a-4321-902c-12a52a2fe4da","Type":"ContainerStarted","Data":"950e8557a995d33501dec28f8d03b14d6d51f1278c1eeab2e2d3e4e6739127c4"} Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.107940 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4nz7s"] Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.109124 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.110925 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.119877 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4nz7s"] Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.242692 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7b69f1c-8ca8-472d-9357-0350e8d732d4-catalog-content\") pod \"certified-operators-4nz7s\" (UID: \"d7b69f1c-8ca8-472d-9357-0350e8d732d4\") " pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.242775 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5jmh\" (UniqueName: \"kubernetes.io/projected/d7b69f1c-8ca8-472d-9357-0350e8d732d4-kube-api-access-b5jmh\") pod \"certified-operators-4nz7s\" (UID: \"d7b69f1c-8ca8-472d-9357-0350e8d732d4\") " pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.242825 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7b69f1c-8ca8-472d-9357-0350e8d732d4-utilities\") pod \"certified-operators-4nz7s\" (UID: \"d7b69f1c-8ca8-472d-9357-0350e8d732d4\") " pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.344892 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5jmh\" (UniqueName: \"kubernetes.io/projected/d7b69f1c-8ca8-472d-9357-0350e8d732d4-kube-api-access-b5jmh\") pod \"certified-operators-4nz7s\" (UID: \"d7b69f1c-8ca8-472d-9357-0350e8d732d4\") " pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.345003 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7b69f1c-8ca8-472d-9357-0350e8d732d4-utilities\") pod \"certified-operators-4nz7s\" (UID: \"d7b69f1c-8ca8-472d-9357-0350e8d732d4\") " pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.345115 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7b69f1c-8ca8-472d-9357-0350e8d732d4-catalog-content\") pod \"certified-operators-4nz7s\" (UID: \"d7b69f1c-8ca8-472d-9357-0350e8d732d4\") " pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.345612 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7b69f1c-8ca8-472d-9357-0350e8d732d4-utilities\") pod \"certified-operators-4nz7s\" (UID: \"d7b69f1c-8ca8-472d-9357-0350e8d732d4\") " pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.345620 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7b69f1c-8ca8-472d-9357-0350e8d732d4-catalog-content\") pod \"certified-operators-4nz7s\" (UID: \"d7b69f1c-8ca8-472d-9357-0350e8d732d4\") " pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.378697 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5jmh\" (UniqueName: \"kubernetes.io/projected/d7b69f1c-8ca8-472d-9357-0350e8d732d4-kube-api-access-b5jmh\") pod \"certified-operators-4nz7s\" (UID: \"d7b69f1c-8ca8-472d-9357-0350e8d732d4\") " pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.441802 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.709419 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-wj4c9" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.721852 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v4nmp"] Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.722872 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v4nmp" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.725120 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.725482 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v4nmp"] Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.805822 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9ws8"] Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.854010 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5452c803-ae28-4bd0-a404-5f3d707b14d5-utilities\") pod \"community-operators-v4nmp\" (UID: \"5452c803-ae28-4bd0-a404-5f3d707b14d5\") " pod="openshift-marketplace/community-operators-v4nmp" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.854059 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5452c803-ae28-4bd0-a404-5f3d707b14d5-catalog-content\") pod \"community-operators-v4nmp\" (UID: \"5452c803-ae28-4bd0-a404-5f3d707b14d5\") " pod="openshift-marketplace/community-operators-v4nmp" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.854104 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pdm6\" (UniqueName: \"kubernetes.io/projected/5452c803-ae28-4bd0-a404-5f3d707b14d5-kube-api-access-6pdm6\") pod \"community-operators-v4nmp\" (UID: \"5452c803-ae28-4bd0-a404-5f3d707b14d5\") " pod="openshift-marketplace/community-operators-v4nmp" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.921461 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4nz7s"] Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.954785 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pdm6\" (UniqueName: \"kubernetes.io/projected/5452c803-ae28-4bd0-a404-5f3d707b14d5-kube-api-access-6pdm6\") pod \"community-operators-v4nmp\" (UID: \"5452c803-ae28-4bd0-a404-5f3d707b14d5\") " pod="openshift-marketplace/community-operators-v4nmp" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.954859 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5452c803-ae28-4bd0-a404-5f3d707b14d5-utilities\") pod \"community-operators-v4nmp\" (UID: \"5452c803-ae28-4bd0-a404-5f3d707b14d5\") " pod="openshift-marketplace/community-operators-v4nmp" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.954888 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5452c803-ae28-4bd0-a404-5f3d707b14d5-catalog-content\") pod \"community-operators-v4nmp\" (UID: \"5452c803-ae28-4bd0-a404-5f3d707b14d5\") " pod="openshift-marketplace/community-operators-v4nmp" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.955354 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5452c803-ae28-4bd0-a404-5f3d707b14d5-utilities\") pod \"community-operators-v4nmp\" (UID: \"5452c803-ae28-4bd0-a404-5f3d707b14d5\") " pod="openshift-marketplace/community-operators-v4nmp" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.955581 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5452c803-ae28-4bd0-a404-5f3d707b14d5-catalog-content\") pod \"community-operators-v4nmp\" (UID: \"5452c803-ae28-4bd0-a404-5f3d707b14d5\") " pod="openshift-marketplace/community-operators-v4nmp" Feb 27 16:30:48 crc kubenswrapper[4814]: I0227 16:30:48.981783 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pdm6\" (UniqueName: \"kubernetes.io/projected/5452c803-ae28-4bd0-a404-5f3d707b14d5-kube-api-access-6pdm6\") pod \"community-operators-v4nmp\" (UID: \"5452c803-ae28-4bd0-a404-5f3d707b14d5\") " pod="openshift-marketplace/community-operators-v4nmp" Feb 27 16:30:49 crc kubenswrapper[4814]: I0227 16:30:49.082579 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v4nmp" Feb 27 16:30:49 crc kubenswrapper[4814]: I0227 16:30:49.488740 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v4nmp"] Feb 27 16:30:49 crc kubenswrapper[4814]: W0227 16:30:49.493974 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5452c803_ae28_4bd0_a404_5f3d707b14d5.slice/crio-0c26e8d3337f729eb3c57f06820ef89627247b1a3706a0fc1f6191ca8520e5cf WatchSource:0}: Error finding container 0c26e8d3337f729eb3c57f06820ef89627247b1a3706a0fc1f6191ca8520e5cf: Status 404 returned error can't find the container with id 0c26e8d3337f729eb3c57f06820ef89627247b1a3706a0fc1f6191ca8520e5cf Feb 27 16:30:49 crc kubenswrapper[4814]: I0227 16:30:49.795973 4814 generic.go:334] "Generic (PLEG): container finished" podID="73643f07-8d1a-4321-902c-12a52a2fe4da" containerID="9c25c6e2a39dfcac4a6c46de2414500a686ad0d301534a0a458e6046b792d9fa" exitCode=0 Feb 27 16:30:49 crc kubenswrapper[4814]: I0227 16:30:49.796136 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5rrh" event={"ID":"73643f07-8d1a-4321-902c-12a52a2fe4da","Type":"ContainerDied","Data":"9c25c6e2a39dfcac4a6c46de2414500a686ad0d301534a0a458e6046b792d9fa"} Feb 27 16:30:49 crc kubenswrapper[4814]: I0227 16:30:49.798552 4814 generic.go:334] "Generic (PLEG): container finished" podID="5452c803-ae28-4bd0-a404-5f3d707b14d5" containerID="dfe433123d7d7574d788f923aae2e15897ea247bb9632c35ce4158ee33303b22" exitCode=0 Feb 27 16:30:49 crc kubenswrapper[4814]: I0227 16:30:49.798619 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4nmp" event={"ID":"5452c803-ae28-4bd0-a404-5f3d707b14d5","Type":"ContainerDied","Data":"dfe433123d7d7574d788f923aae2e15897ea247bb9632c35ce4158ee33303b22"} Feb 27 16:30:49 crc kubenswrapper[4814]: I0227 16:30:49.798695 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4nmp" event={"ID":"5452c803-ae28-4bd0-a404-5f3d707b14d5","Type":"ContainerStarted","Data":"0c26e8d3337f729eb3c57f06820ef89627247b1a3706a0fc1f6191ca8520e5cf"} Feb 27 16:30:49 crc kubenswrapper[4814]: I0227 16:30:49.803521 4814 generic.go:334] "Generic (PLEG): container finished" podID="d7b69f1c-8ca8-472d-9357-0350e8d732d4" containerID="e00b746a40de0797ed3a69e7c1a4f0490c4dd159876129fbebba487a5a115ccb" exitCode=0 Feb 27 16:30:49 crc kubenswrapper[4814]: I0227 16:30:49.803578 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nz7s" event={"ID":"d7b69f1c-8ca8-472d-9357-0350e8d732d4","Type":"ContainerDied","Data":"e00b746a40de0797ed3a69e7c1a4f0490c4dd159876129fbebba487a5a115ccb"} Feb 27 16:30:49 crc kubenswrapper[4814]: I0227 16:30:49.803615 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nz7s" event={"ID":"d7b69f1c-8ca8-472d-9357-0350e8d732d4","Type":"ContainerStarted","Data":"eb96402b588c054e595670e7ac11f855ac8375258207dc2e8d9843a896f08ff6"} Feb 27 16:30:50 crc kubenswrapper[4814]: I0227 16:30:50.516771 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fdwvd"] Feb 27 16:30:50 crc kubenswrapper[4814]: I0227 16:30:50.517920 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fdwvd" Feb 27 16:30:50 crc kubenswrapper[4814]: I0227 16:30:50.521297 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 27 16:30:50 crc kubenswrapper[4814]: I0227 16:30:50.537281 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fdwvd"] Feb 27 16:30:50 crc kubenswrapper[4814]: I0227 16:30:50.690636 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44bnx\" (UniqueName: \"kubernetes.io/projected/01865c94-699f-4fe9-8427-bfe3a2788531-kube-api-access-44bnx\") pod \"redhat-marketplace-fdwvd\" (UID: \"01865c94-699f-4fe9-8427-bfe3a2788531\") " pod="openshift-marketplace/redhat-marketplace-fdwvd" Feb 27 16:30:50 crc kubenswrapper[4814]: I0227 16:30:50.690699 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01865c94-699f-4fe9-8427-bfe3a2788531-utilities\") pod \"redhat-marketplace-fdwvd\" (UID: \"01865c94-699f-4fe9-8427-bfe3a2788531\") " pod="openshift-marketplace/redhat-marketplace-fdwvd" Feb 27 16:30:50 crc kubenswrapper[4814]: I0227 16:30:50.690784 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01865c94-699f-4fe9-8427-bfe3a2788531-catalog-content\") pod \"redhat-marketplace-fdwvd\" (UID: \"01865c94-699f-4fe9-8427-bfe3a2788531\") " pod="openshift-marketplace/redhat-marketplace-fdwvd" Feb 27 16:30:50 crc kubenswrapper[4814]: I0227 16:30:50.795210 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44bnx\" (UniqueName: \"kubernetes.io/projected/01865c94-699f-4fe9-8427-bfe3a2788531-kube-api-access-44bnx\") pod \"redhat-marketplace-fdwvd\" (UID: \"01865c94-699f-4fe9-8427-bfe3a2788531\") " pod="openshift-marketplace/redhat-marketplace-fdwvd" Feb 27 16:30:50 crc kubenswrapper[4814]: I0227 16:30:50.795292 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01865c94-699f-4fe9-8427-bfe3a2788531-utilities\") pod \"redhat-marketplace-fdwvd\" (UID: \"01865c94-699f-4fe9-8427-bfe3a2788531\") " pod="openshift-marketplace/redhat-marketplace-fdwvd" Feb 27 16:30:50 crc kubenswrapper[4814]: I0227 16:30:50.795416 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01865c94-699f-4fe9-8427-bfe3a2788531-catalog-content\") pod \"redhat-marketplace-fdwvd\" (UID: \"01865c94-699f-4fe9-8427-bfe3a2788531\") " pod="openshift-marketplace/redhat-marketplace-fdwvd" Feb 27 16:30:50 crc kubenswrapper[4814]: I0227 16:30:50.796018 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01865c94-699f-4fe9-8427-bfe3a2788531-catalog-content\") pod \"redhat-marketplace-fdwvd\" (UID: \"01865c94-699f-4fe9-8427-bfe3a2788531\") " pod="openshift-marketplace/redhat-marketplace-fdwvd" Feb 27 16:30:50 crc kubenswrapper[4814]: I0227 16:30:50.796239 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01865c94-699f-4fe9-8427-bfe3a2788531-utilities\") pod \"redhat-marketplace-fdwvd\" (UID: \"01865c94-699f-4fe9-8427-bfe3a2788531\") " pod="openshift-marketplace/redhat-marketplace-fdwvd" Feb 27 16:30:50 crc kubenswrapper[4814]: I0227 16:30:50.813032 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5rrh" event={"ID":"73643f07-8d1a-4321-902c-12a52a2fe4da","Type":"ContainerStarted","Data":"d0db7181d5ee7b26e3303193ab1566306ba546ce39825ba44c7ca3070332fb67"} Feb 27 16:30:50 crc kubenswrapper[4814]: I0227 16:30:50.828612 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44bnx\" (UniqueName: \"kubernetes.io/projected/01865c94-699f-4fe9-8427-bfe3a2788531-kube-api-access-44bnx\") pod \"redhat-marketplace-fdwvd\" (UID: \"01865c94-699f-4fe9-8427-bfe3a2788531\") " pod="openshift-marketplace/redhat-marketplace-fdwvd" Feb 27 16:30:50 crc kubenswrapper[4814]: I0227 16:30:50.864481 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r5rrh" podStartSLOduration=2.084304542 podStartE2EDuration="4.864455228s" podCreationTimestamp="2026-02-27 16:30:46 +0000 UTC" firstStartedPulling="2026-02-27 16:30:47.738974313 +0000 UTC m=+460.191599143" lastFinishedPulling="2026-02-27 16:30:50.519124959 +0000 UTC m=+462.971749829" observedRunningTime="2026-02-27 16:30:50.860973265 +0000 UTC m=+463.313598135" watchObservedRunningTime="2026-02-27 16:30:50.864455228 +0000 UTC m=+463.317080098" Feb 27 16:30:50 crc kubenswrapper[4814]: I0227 16:30:50.872024 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fdwvd" Feb 27 16:30:51 crc kubenswrapper[4814]: I0227 16:30:51.427163 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fdwvd"] Feb 27 16:30:51 crc kubenswrapper[4814]: I0227 16:30:51.840960 4814 generic.go:334] "Generic (PLEG): container finished" podID="01865c94-699f-4fe9-8427-bfe3a2788531" containerID="98ecabc73b0cc7ad451bb2335d00368eb2f111b9f5303ce173b5d3f72148bb68" exitCode=0 Feb 27 16:30:51 crc kubenswrapper[4814]: I0227 16:30:51.841643 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdwvd" event={"ID":"01865c94-699f-4fe9-8427-bfe3a2788531","Type":"ContainerDied","Data":"98ecabc73b0cc7ad451bb2335d00368eb2f111b9f5303ce173b5d3f72148bb68"} Feb 27 16:30:51 crc kubenswrapper[4814]: I0227 16:30:51.841711 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdwvd" event={"ID":"01865c94-699f-4fe9-8427-bfe3a2788531","Type":"ContainerStarted","Data":"300487b179d5c1c48fb328598acdad7f95b2aca08d09009c857f53400747081c"} Feb 27 16:30:52 crc kubenswrapper[4814]: I0227 16:30:52.902444 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:30:52 crc kubenswrapper[4814]: I0227 16:30:52.902883 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:30:53 crc kubenswrapper[4814]: E0227 16:30:53.161047 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/NetworkManager-dispatcher.service\": RecentStats: unable to find data in memory cache]" Feb 27 16:30:53 crc kubenswrapper[4814]: I0227 16:30:53.859133 4814 generic.go:334] "Generic (PLEG): container finished" podID="d7b69f1c-8ca8-472d-9357-0350e8d732d4" containerID="cee7dd070e9a8af352e007ab238c67b868a4737799d1e7e446b2b08fe9956b85" exitCode=0 Feb 27 16:30:53 crc kubenswrapper[4814]: I0227 16:30:53.859232 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nz7s" event={"ID":"d7b69f1c-8ca8-472d-9357-0350e8d732d4","Type":"ContainerDied","Data":"cee7dd070e9a8af352e007ab238c67b868a4737799d1e7e446b2b08fe9956b85"} Feb 27 16:30:53 crc kubenswrapper[4814]: I0227 16:30:53.864785 4814 generic.go:334] "Generic (PLEG): container finished" podID="01865c94-699f-4fe9-8427-bfe3a2788531" containerID="c2ffcaedcba4d62136875c68a3182643922ea4a27a0d0209f6d7167ef90d1268" exitCode=0 Feb 27 16:30:53 crc kubenswrapper[4814]: I0227 16:30:53.864998 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdwvd" event={"ID":"01865c94-699f-4fe9-8427-bfe3a2788531","Type":"ContainerDied","Data":"c2ffcaedcba4d62136875c68a3182643922ea4a27a0d0209f6d7167ef90d1268"} Feb 27 16:30:53 crc kubenswrapper[4814]: I0227 16:30:53.869046 4814 generic.go:334] "Generic (PLEG): container finished" podID="5452c803-ae28-4bd0-a404-5f3d707b14d5" containerID="07df7f0f0d897c49cbec2f53833dba3b1ad78c9ba23a60b6fcef7591d60c77c9" exitCode=0 Feb 27 16:30:53 crc kubenswrapper[4814]: I0227 16:30:53.869086 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4nmp" event={"ID":"5452c803-ae28-4bd0-a404-5f3d707b14d5","Type":"ContainerDied","Data":"07df7f0f0d897c49cbec2f53833dba3b1ad78c9ba23a60b6fcef7591d60c77c9"} Feb 27 16:30:54 crc kubenswrapper[4814]: I0227 16:30:54.878231 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4nmp" event={"ID":"5452c803-ae28-4bd0-a404-5f3d707b14d5","Type":"ContainerStarted","Data":"af513b1160b8bba968478935ef6eec66af4ad2f950eb6b3613463448d796839a"} Feb 27 16:30:54 crc kubenswrapper[4814]: I0227 16:30:54.899978 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v4nmp" podStartSLOduration=2.405044106 podStartE2EDuration="6.899962591s" podCreationTimestamp="2026-02-27 16:30:48 +0000 UTC" firstStartedPulling="2026-02-27 16:30:49.802181831 +0000 UTC m=+462.254806701" lastFinishedPulling="2026-02-27 16:30:54.297100326 +0000 UTC m=+466.749725186" observedRunningTime="2026-02-27 16:30:54.896462498 +0000 UTC m=+467.349087348" watchObservedRunningTime="2026-02-27 16:30:54.899962591 +0000 UTC m=+467.352587421" Feb 27 16:30:55 crc kubenswrapper[4814]: I0227 16:30:55.888911 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fdwvd" event={"ID":"01865c94-699f-4fe9-8427-bfe3a2788531","Type":"ContainerStarted","Data":"fe782aaaaecf273d53a1efeacc1947518705c7a91315a655ad4d8cedb3869143"} Feb 27 16:30:55 crc kubenswrapper[4814]: I0227 16:30:55.892043 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nz7s" event={"ID":"d7b69f1c-8ca8-472d-9357-0350e8d732d4","Type":"ContainerStarted","Data":"e13d476b1802f964dc3a86730c1f816149b7d18cae7d6acb6521cb0c64817f4a"} Feb 27 16:30:55 crc kubenswrapper[4814]: I0227 16:30:55.914014 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fdwvd" podStartSLOduration=2.936474991 podStartE2EDuration="5.913994596s" podCreationTimestamp="2026-02-27 16:30:50 +0000 UTC" firstStartedPulling="2026-02-27 16:30:51.844168742 +0000 UTC m=+464.296793582" lastFinishedPulling="2026-02-27 16:30:54.821688347 +0000 UTC m=+467.274313187" observedRunningTime="2026-02-27 16:30:55.911268778 +0000 UTC m=+468.363893608" watchObservedRunningTime="2026-02-27 16:30:55.913994596 +0000 UTC m=+468.366619426" Feb 27 16:30:55 crc kubenswrapper[4814]: I0227 16:30:55.931209 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4nz7s" podStartSLOduration=2.968477615 podStartE2EDuration="7.931190903s" podCreationTimestamp="2026-02-27 16:30:48 +0000 UTC" firstStartedPulling="2026-02-27 16:30:49.805465737 +0000 UTC m=+462.258090597" lastFinishedPulling="2026-02-27 16:30:54.768179015 +0000 UTC m=+467.220803885" observedRunningTime="2026-02-27 16:30:55.927165043 +0000 UTC m=+468.379789873" watchObservedRunningTime="2026-02-27 16:30:55.931190903 +0000 UTC m=+468.383815723" Feb 27 16:30:56 crc kubenswrapper[4814]: I0227 16:30:56.660597 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r5rrh" Feb 27 16:30:56 crc kubenswrapper[4814]: I0227 16:30:56.660987 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r5rrh" Feb 27 16:30:57 crc kubenswrapper[4814]: I0227 16:30:57.702306 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-r5rrh" podUID="73643f07-8d1a-4321-902c-12a52a2fe4da" containerName="registry-server" probeResult="failure" output=< Feb 27 16:30:57 crc kubenswrapper[4814]: timeout: failed to connect service ":50051" within 1s Feb 27 16:30:57 crc kubenswrapper[4814]: > Feb 27 16:30:58 crc kubenswrapper[4814]: I0227 16:30:58.442762 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 16:30:58 crc kubenswrapper[4814]: I0227 16:30:58.442840 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 16:30:58 crc kubenswrapper[4814]: I0227 16:30:58.502035 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 16:30:59 crc kubenswrapper[4814]: I0227 16:30:59.084637 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v4nmp" Feb 27 16:30:59 crc kubenswrapper[4814]: I0227 16:30:59.084708 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v4nmp" Feb 27 16:30:59 crc kubenswrapper[4814]: I0227 16:30:59.155951 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v4nmp" Feb 27 16:30:59 crc kubenswrapper[4814]: I0227 16:30:59.968406 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v4nmp" Feb 27 16:31:00 crc kubenswrapper[4814]: I0227 16:31:00.873879 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fdwvd" Feb 27 16:31:00 crc kubenswrapper[4814]: I0227 16:31:00.873964 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fdwvd" Feb 27 16:31:00 crc kubenswrapper[4814]: I0227 16:31:00.930520 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fdwvd" Feb 27 16:31:01 crc kubenswrapper[4814]: I0227 16:31:01.019532 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fdwvd" Feb 27 16:31:06 crc kubenswrapper[4814]: I0227 16:31:06.736928 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r5rrh" Feb 27 16:31:06 crc kubenswrapper[4814]: I0227 16:31:06.820825 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r5rrh" Feb 27 16:31:08 crc kubenswrapper[4814]: I0227 16:31:08.516239 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 16:31:13 crc kubenswrapper[4814]: I0227 16:31:13.858540 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" podUID="2b4a4d02-59a6-4f74-a4cc-2154d1115361" containerName="registry" containerID="cri-o://e3d9c465ec3deb6bf096badcd6f47ff7aab7c32dd2b06a6b8d5f07b4efac3a8a" gracePeriod=30 Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.036573 4814 generic.go:334] "Generic (PLEG): container finished" podID="2b4a4d02-59a6-4f74-a4cc-2154d1115361" containerID="e3d9c465ec3deb6bf096badcd6f47ff7aab7c32dd2b06a6b8d5f07b4efac3a8a" exitCode=0 Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.036636 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" event={"ID":"2b4a4d02-59a6-4f74-a4cc-2154d1115361","Type":"ContainerDied","Data":"e3d9c465ec3deb6bf096badcd6f47ff7aab7c32dd2b06a6b8d5f07b4efac3a8a"} Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.355928 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.492760 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.492961 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b4a4d02-59a6-4f74-a4cc-2154d1115361-registry-certificates\") pod \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.492990 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b4a4d02-59a6-4f74-a4cc-2154d1115361-trusted-ca\") pod \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.493044 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7cqz\" (UniqueName: \"kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-kube-api-access-m7cqz\") pod \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.493155 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b4a4d02-59a6-4f74-a4cc-2154d1115361-installation-pull-secrets\") pod \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.493284 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-registry-tls\") pod \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.493357 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b4a4d02-59a6-4f74-a4cc-2154d1115361-ca-trust-extracted\") pod \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.493382 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-bound-sa-token\") pod \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\" (UID: \"2b4a4d02-59a6-4f74-a4cc-2154d1115361\") " Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.494241 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b4a4d02-59a6-4f74-a4cc-2154d1115361-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "2b4a4d02-59a6-4f74-a4cc-2154d1115361" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.494460 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b4a4d02-59a6-4f74-a4cc-2154d1115361-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "2b4a4d02-59a6-4f74-a4cc-2154d1115361" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.501476 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "2b4a4d02-59a6-4f74-a4cc-2154d1115361" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.502818 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b4a4d02-59a6-4f74-a4cc-2154d1115361-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "2b4a4d02-59a6-4f74-a4cc-2154d1115361" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.502918 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-kube-api-access-m7cqz" (OuterVolumeSpecName: "kube-api-access-m7cqz") pod "2b4a4d02-59a6-4f74-a4cc-2154d1115361" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361"). InnerVolumeSpecName "kube-api-access-m7cqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.509116 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "2b4a4d02-59a6-4f74-a4cc-2154d1115361" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.511003 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "2b4a4d02-59a6-4f74-a4cc-2154d1115361" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.521036 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b4a4d02-59a6-4f74-a4cc-2154d1115361-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "2b4a4d02-59a6-4f74-a4cc-2154d1115361" (UID: "2b4a4d02-59a6-4f74-a4cc-2154d1115361"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.594528 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7cqz\" (UniqueName: \"kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-kube-api-access-m7cqz\") on node \"crc\" DevicePath \"\"" Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.594575 4814 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b4a4d02-59a6-4f74-a4cc-2154d1115361-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.594589 4814 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.594601 4814 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b4a4d02-59a6-4f74-a4cc-2154d1115361-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.594613 4814 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b4a4d02-59a6-4f74-a4cc-2154d1115361-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.594623 4814 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b4a4d02-59a6-4f74-a4cc-2154d1115361-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 27 16:31:14 crc kubenswrapper[4814]: I0227 16:31:14.594632 4814 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b4a4d02-59a6-4f74-a4cc-2154d1115361-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:31:15 crc kubenswrapper[4814]: I0227 16:31:15.046186 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" event={"ID":"2b4a4d02-59a6-4f74-a4cc-2154d1115361","Type":"ContainerDied","Data":"fd04451f58791f8461df970d3d11dff3a454c2366400d4f43cffdb7695978e44"} Feb 27 16:31:15 crc kubenswrapper[4814]: I0227 16:31:15.046274 4814 scope.go:117] "RemoveContainer" containerID="e3d9c465ec3deb6bf096badcd6f47ff7aab7c32dd2b06a6b8d5f07b4efac3a8a" Feb 27 16:31:15 crc kubenswrapper[4814]: I0227 16:31:15.046286 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-q9ws8" Feb 27 16:31:15 crc kubenswrapper[4814]: I0227 16:31:15.077761 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9ws8"] Feb 27 16:31:15 crc kubenswrapper[4814]: I0227 16:31:15.082710 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9ws8"] Feb 27 16:31:16 crc kubenswrapper[4814]: I0227 16:31:16.501763 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b4a4d02-59a6-4f74-a4cc-2154d1115361" path="/var/lib/kubelet/pods/2b4a4d02-59a6-4f74-a4cc-2154d1115361/volumes" Feb 27 16:31:22 crc kubenswrapper[4814]: I0227 16:31:22.902876 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:31:22 crc kubenswrapper[4814]: I0227 16:31:22.903772 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:31:52 crc kubenswrapper[4814]: I0227 16:31:52.907625 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:31:52 crc kubenswrapper[4814]: I0227 16:31:52.908631 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:31:52 crc kubenswrapper[4814]: I0227 16:31:52.908713 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:31:52 crc kubenswrapper[4814]: I0227 16:31:52.909728 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7d5c0ac92b2e6bacf90ee7b6e5821df1c3f5a2b33024a13ffeb8403e774e5521"} pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 16:31:52 crc kubenswrapper[4814]: I0227 16:31:52.909871 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" containerID="cri-o://7d5c0ac92b2e6bacf90ee7b6e5821df1c3f5a2b33024a13ffeb8403e774e5521" gracePeriod=600 Feb 27 16:31:53 crc kubenswrapper[4814]: I0227 16:31:53.320933 4814 generic.go:334] "Generic (PLEG): container finished" podID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerID="7d5c0ac92b2e6bacf90ee7b6e5821df1c3f5a2b33024a13ffeb8403e774e5521" exitCode=0 Feb 27 16:31:53 crc kubenswrapper[4814]: I0227 16:31:53.321006 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerDied","Data":"7d5c0ac92b2e6bacf90ee7b6e5821df1c3f5a2b33024a13ffeb8403e774e5521"} Feb 27 16:31:53 crc kubenswrapper[4814]: I0227 16:31:53.321222 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"5179547dd473be027ccf5a04ad384d8f4660f7e3ba2b8851cd7be0f851fdbb8a"} Feb 27 16:31:53 crc kubenswrapper[4814]: I0227 16:31:53.321239 4814 scope.go:117] "RemoveContainer" containerID="c71a3e11a4dc0f7883e443fd1e5cc111ec461bd4a62ee855e618f53fda86c223" Feb 27 16:32:00 crc kubenswrapper[4814]: I0227 16:32:00.155726 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536832-blzjj"] Feb 27 16:32:00 crc kubenswrapper[4814]: E0227 16:32:00.159462 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b4a4d02-59a6-4f74-a4cc-2154d1115361" containerName="registry" Feb 27 16:32:00 crc kubenswrapper[4814]: I0227 16:32:00.159683 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b4a4d02-59a6-4f74-a4cc-2154d1115361" containerName="registry" Feb 27 16:32:00 crc kubenswrapper[4814]: I0227 16:32:00.160096 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b4a4d02-59a6-4f74-a4cc-2154d1115361" containerName="registry" Feb 27 16:32:00 crc kubenswrapper[4814]: I0227 16:32:00.161406 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536832-blzjj" Feb 27 16:32:00 crc kubenswrapper[4814]: I0227 16:32:00.166529 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:32:00 crc kubenswrapper[4814]: I0227 16:32:00.166854 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 16:32:00 crc kubenswrapper[4814]: I0227 16:32:00.169600 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:32:00 crc kubenswrapper[4814]: I0227 16:32:00.171902 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536832-blzjj"] Feb 27 16:32:00 crc kubenswrapper[4814]: I0227 16:32:00.279851 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjf5w\" (UniqueName: \"kubernetes.io/projected/509692fb-2332-42cc-a678-b447aa1ed437-kube-api-access-bjf5w\") pod \"auto-csr-approver-29536832-blzjj\" (UID: \"509692fb-2332-42cc-a678-b447aa1ed437\") " pod="openshift-infra/auto-csr-approver-29536832-blzjj" Feb 27 16:32:00 crc kubenswrapper[4814]: I0227 16:32:00.381563 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjf5w\" (UniqueName: \"kubernetes.io/projected/509692fb-2332-42cc-a678-b447aa1ed437-kube-api-access-bjf5w\") pod \"auto-csr-approver-29536832-blzjj\" (UID: \"509692fb-2332-42cc-a678-b447aa1ed437\") " pod="openshift-infra/auto-csr-approver-29536832-blzjj" Feb 27 16:32:00 crc kubenswrapper[4814]: I0227 16:32:00.421200 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjf5w\" (UniqueName: \"kubernetes.io/projected/509692fb-2332-42cc-a678-b447aa1ed437-kube-api-access-bjf5w\") pod \"auto-csr-approver-29536832-blzjj\" (UID: \"509692fb-2332-42cc-a678-b447aa1ed437\") " pod="openshift-infra/auto-csr-approver-29536832-blzjj" Feb 27 16:32:00 crc kubenswrapper[4814]: I0227 16:32:00.496527 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536832-blzjj" Feb 27 16:32:00 crc kubenswrapper[4814]: I0227 16:32:00.777466 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536832-blzjj"] Feb 27 16:32:00 crc kubenswrapper[4814]: I0227 16:32:00.788714 4814 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 16:32:01 crc kubenswrapper[4814]: I0227 16:32:01.385394 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536832-blzjj" event={"ID":"509692fb-2332-42cc-a678-b447aa1ed437","Type":"ContainerStarted","Data":"9a076f3d54f26d149de43f65672eab8cddc60092292b01988f1c667f9dfa51fe"} Feb 27 16:32:02 crc kubenswrapper[4814]: I0227 16:32:02.397934 4814 generic.go:334] "Generic (PLEG): container finished" podID="509692fb-2332-42cc-a678-b447aa1ed437" containerID="64aab1a2f7b2095ddb3c933a98f950873b3fa8b4b3a3564427139a7ffa933e44" exitCode=0 Feb 27 16:32:02 crc kubenswrapper[4814]: I0227 16:32:02.398032 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536832-blzjj" event={"ID":"509692fb-2332-42cc-a678-b447aa1ed437","Type":"ContainerDied","Data":"64aab1a2f7b2095ddb3c933a98f950873b3fa8b4b3a3564427139a7ffa933e44"} Feb 27 16:32:03 crc kubenswrapper[4814]: I0227 16:32:03.778030 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536832-blzjj" Feb 27 16:32:03 crc kubenswrapper[4814]: I0227 16:32:03.840877 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjf5w\" (UniqueName: \"kubernetes.io/projected/509692fb-2332-42cc-a678-b447aa1ed437-kube-api-access-bjf5w\") pod \"509692fb-2332-42cc-a678-b447aa1ed437\" (UID: \"509692fb-2332-42cc-a678-b447aa1ed437\") " Feb 27 16:32:03 crc kubenswrapper[4814]: I0227 16:32:03.851656 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/509692fb-2332-42cc-a678-b447aa1ed437-kube-api-access-bjf5w" (OuterVolumeSpecName: "kube-api-access-bjf5w") pod "509692fb-2332-42cc-a678-b447aa1ed437" (UID: "509692fb-2332-42cc-a678-b447aa1ed437"). InnerVolumeSpecName "kube-api-access-bjf5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:32:03 crc kubenswrapper[4814]: I0227 16:32:03.943209 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjf5w\" (UniqueName: \"kubernetes.io/projected/509692fb-2332-42cc-a678-b447aa1ed437-kube-api-access-bjf5w\") on node \"crc\" DevicePath \"\"" Feb 27 16:32:04 crc kubenswrapper[4814]: I0227 16:32:04.417203 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536832-blzjj" event={"ID":"509692fb-2332-42cc-a678-b447aa1ed437","Type":"ContainerDied","Data":"9a076f3d54f26d149de43f65672eab8cddc60092292b01988f1c667f9dfa51fe"} Feb 27 16:32:04 crc kubenswrapper[4814]: I0227 16:32:04.417295 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536832-blzjj" Feb 27 16:32:04 crc kubenswrapper[4814]: I0227 16:32:04.417322 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a076f3d54f26d149de43f65672eab8cddc60092292b01988f1c667f9dfa51fe" Feb 27 16:32:04 crc kubenswrapper[4814]: I0227 16:32:04.881935 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536826-dgwn4"] Feb 27 16:32:04 crc kubenswrapper[4814]: I0227 16:32:04.894158 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536826-dgwn4"] Feb 27 16:32:06 crc kubenswrapper[4814]: I0227 16:32:06.503050 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95913cb4-39f1-44c0-ac49-0a2d51047679" path="/var/lib/kubelet/pods/95913cb4-39f1-44c0-ac49-0a2d51047679/volumes" Feb 27 16:33:25 crc kubenswrapper[4814]: I0227 16:33:25.715895 4814 scope.go:117] "RemoveContainer" containerID="3ecfe545b52140b6f8d8e2da63d6dd1c7dbf2fe5f9fed179b29c2932932e24e2" Feb 27 16:33:25 crc kubenswrapper[4814]: I0227 16:33:25.748640 4814 scope.go:117] "RemoveContainer" containerID="526183fdafe1c67c9473802ba8555164094bdf32a27178105649c4bb6176725b" Feb 27 16:34:00 crc kubenswrapper[4814]: I0227 16:34:00.146939 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536834-w2h2s"] Feb 27 16:34:00 crc kubenswrapper[4814]: E0227 16:34:00.149320 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="509692fb-2332-42cc-a678-b447aa1ed437" containerName="oc" Feb 27 16:34:00 crc kubenswrapper[4814]: I0227 16:34:00.149422 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="509692fb-2332-42cc-a678-b447aa1ed437" containerName="oc" Feb 27 16:34:00 crc kubenswrapper[4814]: I0227 16:34:00.149708 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="509692fb-2332-42cc-a678-b447aa1ed437" containerName="oc" Feb 27 16:34:00 crc kubenswrapper[4814]: I0227 16:34:00.150318 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536834-w2h2s" Feb 27 16:34:00 crc kubenswrapper[4814]: I0227 16:34:00.153351 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 16:34:00 crc kubenswrapper[4814]: I0227 16:34:00.153677 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:34:00 crc kubenswrapper[4814]: I0227 16:34:00.154560 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:34:00 crc kubenswrapper[4814]: I0227 16:34:00.163829 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536834-w2h2s"] Feb 27 16:34:00 crc kubenswrapper[4814]: I0227 16:34:00.228473 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlvgk\" (UniqueName: \"kubernetes.io/projected/d7ca68cb-43ee-405b-a9bf-8a6dc109ee27-kube-api-access-dlvgk\") pod \"auto-csr-approver-29536834-w2h2s\" (UID: \"d7ca68cb-43ee-405b-a9bf-8a6dc109ee27\") " pod="openshift-infra/auto-csr-approver-29536834-w2h2s" Feb 27 16:34:00 crc kubenswrapper[4814]: I0227 16:34:00.330685 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlvgk\" (UniqueName: \"kubernetes.io/projected/d7ca68cb-43ee-405b-a9bf-8a6dc109ee27-kube-api-access-dlvgk\") pod \"auto-csr-approver-29536834-w2h2s\" (UID: \"d7ca68cb-43ee-405b-a9bf-8a6dc109ee27\") " pod="openshift-infra/auto-csr-approver-29536834-w2h2s" Feb 27 16:34:00 crc kubenswrapper[4814]: I0227 16:34:00.355469 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlvgk\" (UniqueName: \"kubernetes.io/projected/d7ca68cb-43ee-405b-a9bf-8a6dc109ee27-kube-api-access-dlvgk\") pod \"auto-csr-approver-29536834-w2h2s\" (UID: \"d7ca68cb-43ee-405b-a9bf-8a6dc109ee27\") " pod="openshift-infra/auto-csr-approver-29536834-w2h2s" Feb 27 16:34:00 crc kubenswrapper[4814]: I0227 16:34:00.474460 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536834-w2h2s" Feb 27 16:34:00 crc kubenswrapper[4814]: I0227 16:34:00.743046 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536834-w2h2s"] Feb 27 16:34:01 crc kubenswrapper[4814]: I0227 16:34:01.311124 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536834-w2h2s" event={"ID":"d7ca68cb-43ee-405b-a9bf-8a6dc109ee27","Type":"ContainerStarted","Data":"c7f0233c0dc4f95bee12ec6b4cf1657d4156e55e5514cbfbce14635cb9884d05"} Feb 27 16:34:02 crc kubenswrapper[4814]: I0227 16:34:02.327393 4814 generic.go:334] "Generic (PLEG): container finished" podID="d7ca68cb-43ee-405b-a9bf-8a6dc109ee27" containerID="937fa75707ae28a50ee5d55cfc7a9b9dfa99e3bc62026ff27ce4d0a8f6c024d6" exitCode=0 Feb 27 16:34:02 crc kubenswrapper[4814]: I0227 16:34:02.327506 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536834-w2h2s" event={"ID":"d7ca68cb-43ee-405b-a9bf-8a6dc109ee27","Type":"ContainerDied","Data":"937fa75707ae28a50ee5d55cfc7a9b9dfa99e3bc62026ff27ce4d0a8f6c024d6"} Feb 27 16:34:03 crc kubenswrapper[4814]: I0227 16:34:03.597353 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536834-w2h2s" Feb 27 16:34:03 crc kubenswrapper[4814]: I0227 16:34:03.778418 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlvgk\" (UniqueName: \"kubernetes.io/projected/d7ca68cb-43ee-405b-a9bf-8a6dc109ee27-kube-api-access-dlvgk\") pod \"d7ca68cb-43ee-405b-a9bf-8a6dc109ee27\" (UID: \"d7ca68cb-43ee-405b-a9bf-8a6dc109ee27\") " Feb 27 16:34:03 crc kubenswrapper[4814]: I0227 16:34:03.785213 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7ca68cb-43ee-405b-a9bf-8a6dc109ee27-kube-api-access-dlvgk" (OuterVolumeSpecName: "kube-api-access-dlvgk") pod "d7ca68cb-43ee-405b-a9bf-8a6dc109ee27" (UID: "d7ca68cb-43ee-405b-a9bf-8a6dc109ee27"). InnerVolumeSpecName "kube-api-access-dlvgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:34:03 crc kubenswrapper[4814]: I0227 16:34:03.880970 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlvgk\" (UniqueName: \"kubernetes.io/projected/d7ca68cb-43ee-405b-a9bf-8a6dc109ee27-kube-api-access-dlvgk\") on node \"crc\" DevicePath \"\"" Feb 27 16:34:04 crc kubenswrapper[4814]: I0227 16:34:04.345112 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536834-w2h2s" event={"ID":"d7ca68cb-43ee-405b-a9bf-8a6dc109ee27","Type":"ContainerDied","Data":"c7f0233c0dc4f95bee12ec6b4cf1657d4156e55e5514cbfbce14635cb9884d05"} Feb 27 16:34:04 crc kubenswrapper[4814]: I0227 16:34:04.345168 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7f0233c0dc4f95bee12ec6b4cf1657d4156e55e5514cbfbce14635cb9884d05" Feb 27 16:34:04 crc kubenswrapper[4814]: I0227 16:34:04.345163 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536834-w2h2s" Feb 27 16:34:04 crc kubenswrapper[4814]: I0227 16:34:04.680876 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536828-g5rhp"] Feb 27 16:34:04 crc kubenswrapper[4814]: I0227 16:34:04.688240 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536828-g5rhp"] Feb 27 16:34:06 crc kubenswrapper[4814]: I0227 16:34:06.498779 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd3bb66c-e36b-456f-8051-509099d63405" path="/var/lib/kubelet/pods/dd3bb66c-e36b-456f-8051-509099d63405/volumes" Feb 27 16:34:22 crc kubenswrapper[4814]: I0227 16:34:22.902991 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:34:22 crc kubenswrapper[4814]: I0227 16:34:22.903711 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:34:25 crc kubenswrapper[4814]: I0227 16:34:25.818733 4814 scope.go:117] "RemoveContainer" containerID="81ef0dcfb1aa0d2d1bba5a21ad84b96dea5ed110e08c21309ab4e1e86a512043" Feb 27 16:34:25 crc kubenswrapper[4814]: I0227 16:34:25.855367 4814 scope.go:117] "RemoveContainer" containerID="d70877336baf2a47b41a630c1f36e75d576381e67a5869647a77e9d1f17e9a43" Feb 27 16:34:25 crc kubenswrapper[4814]: I0227 16:34:25.874585 4814 scope.go:117] "RemoveContainer" containerID="9a038ee1f7a4efc09ec1a2ebfb4250554cbbcea6f81b89bc6e1317e43f091d3c" Feb 27 16:34:25 crc kubenswrapper[4814]: I0227 16:34:25.916767 4814 scope.go:117] "RemoveContainer" containerID="d52b0abb8f8cbbfa2859761e99c665f2fed9ad8b6ff546737caaee892c4d4731" Feb 27 16:34:52 crc kubenswrapper[4814]: I0227 16:34:52.903173 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:34:52 crc kubenswrapper[4814]: I0227 16:34:52.903789 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:35:22 crc kubenswrapper[4814]: I0227 16:35:22.902356 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:35:22 crc kubenswrapper[4814]: I0227 16:35:22.903352 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:35:22 crc kubenswrapper[4814]: I0227 16:35:22.903425 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:35:22 crc kubenswrapper[4814]: I0227 16:35:22.904238 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5179547dd473be027ccf5a04ad384d8f4660f7e3ba2b8851cd7be0f851fdbb8a"} pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 16:35:22 crc kubenswrapper[4814]: I0227 16:35:22.904360 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" containerID="cri-o://5179547dd473be027ccf5a04ad384d8f4660f7e3ba2b8851cd7be0f851fdbb8a" gracePeriod=600 Feb 27 16:35:23 crc kubenswrapper[4814]: I0227 16:35:23.886372 4814 generic.go:334] "Generic (PLEG): container finished" podID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerID="5179547dd473be027ccf5a04ad384d8f4660f7e3ba2b8851cd7be0f851fdbb8a" exitCode=0 Feb 27 16:35:23 crc kubenswrapper[4814]: I0227 16:35:23.886501 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerDied","Data":"5179547dd473be027ccf5a04ad384d8f4660f7e3ba2b8851cd7be0f851fdbb8a"} Feb 27 16:35:23 crc kubenswrapper[4814]: I0227 16:35:23.886864 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"7a221694cd83011f5779a509b7774635f372c6adcfadbeae289b463611af2c58"} Feb 27 16:35:23 crc kubenswrapper[4814]: I0227 16:35:23.886902 4814 scope.go:117] "RemoveContainer" containerID="7d5c0ac92b2e6bacf90ee7b6e5821df1c3f5a2b33024a13ffeb8403e774e5521" Feb 27 16:36:00 crc kubenswrapper[4814]: I0227 16:36:00.153124 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536836-7pdzj"] Feb 27 16:36:00 crc kubenswrapper[4814]: E0227 16:36:00.154072 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ca68cb-43ee-405b-a9bf-8a6dc109ee27" containerName="oc" Feb 27 16:36:00 crc kubenswrapper[4814]: I0227 16:36:00.154095 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ca68cb-43ee-405b-a9bf-8a6dc109ee27" containerName="oc" Feb 27 16:36:00 crc kubenswrapper[4814]: I0227 16:36:00.154245 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7ca68cb-43ee-405b-a9bf-8a6dc109ee27" containerName="oc" Feb 27 16:36:00 crc kubenswrapper[4814]: I0227 16:36:00.154975 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536836-7pdzj" Feb 27 16:36:00 crc kubenswrapper[4814]: I0227 16:36:00.158157 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:36:00 crc kubenswrapper[4814]: I0227 16:36:00.158668 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:36:00 crc kubenswrapper[4814]: I0227 16:36:00.159431 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 16:36:00 crc kubenswrapper[4814]: I0227 16:36:00.175804 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536836-7pdzj"] Feb 27 16:36:00 crc kubenswrapper[4814]: I0227 16:36:00.282858 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xndwx\" (UniqueName: \"kubernetes.io/projected/83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91-kube-api-access-xndwx\") pod \"auto-csr-approver-29536836-7pdzj\" (UID: \"83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91\") " pod="openshift-infra/auto-csr-approver-29536836-7pdzj" Feb 27 16:36:00 crc kubenswrapper[4814]: I0227 16:36:00.384204 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xndwx\" (UniqueName: \"kubernetes.io/projected/83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91-kube-api-access-xndwx\") pod \"auto-csr-approver-29536836-7pdzj\" (UID: \"83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91\") " pod="openshift-infra/auto-csr-approver-29536836-7pdzj" Feb 27 16:36:00 crc kubenswrapper[4814]: I0227 16:36:00.420522 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xndwx\" (UniqueName: \"kubernetes.io/projected/83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91-kube-api-access-xndwx\") pod \"auto-csr-approver-29536836-7pdzj\" (UID: \"83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91\") " pod="openshift-infra/auto-csr-approver-29536836-7pdzj" Feb 27 16:36:00 crc kubenswrapper[4814]: I0227 16:36:00.482771 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536836-7pdzj" Feb 27 16:36:00 crc kubenswrapper[4814]: I0227 16:36:00.713488 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536836-7pdzj"] Feb 27 16:36:01 crc kubenswrapper[4814]: I0227 16:36:01.145390 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536836-7pdzj" event={"ID":"83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91","Type":"ContainerStarted","Data":"fe619277ec30081c7b4d088516a652727509d271a3c1d7dd365ba01bf701d716"} Feb 27 16:36:02 crc kubenswrapper[4814]: I0227 16:36:02.151840 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536836-7pdzj" event={"ID":"83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91","Type":"ContainerStarted","Data":"164ebc8641f4885df5b56586f8ce5d02d3bd9dbba699838e0448b5fbb8d90f12"} Feb 27 16:36:02 crc kubenswrapper[4814]: I0227 16:36:02.169645 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536836-7pdzj" podStartSLOduration=1.21602484 podStartE2EDuration="2.169627008s" podCreationTimestamp="2026-02-27 16:36:00 +0000 UTC" firstStartedPulling="2026-02-27 16:36:00.724549797 +0000 UTC m=+773.177174637" lastFinishedPulling="2026-02-27 16:36:01.678151945 +0000 UTC m=+774.130776805" observedRunningTime="2026-02-27 16:36:02.168580244 +0000 UTC m=+774.621205114" watchObservedRunningTime="2026-02-27 16:36:02.169627008 +0000 UTC m=+774.622251838" Feb 27 16:36:03 crc kubenswrapper[4814]: I0227 16:36:03.161853 4814 generic.go:334] "Generic (PLEG): container finished" podID="83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91" containerID="164ebc8641f4885df5b56586f8ce5d02d3bd9dbba699838e0448b5fbb8d90f12" exitCode=0 Feb 27 16:36:03 crc kubenswrapper[4814]: I0227 16:36:03.162000 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536836-7pdzj" event={"ID":"83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91","Type":"ContainerDied","Data":"164ebc8641f4885df5b56586f8ce5d02d3bd9dbba699838e0448b5fbb8d90f12"} Feb 27 16:36:04 crc kubenswrapper[4814]: I0227 16:36:04.536422 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536836-7pdzj" Feb 27 16:36:04 crc kubenswrapper[4814]: I0227 16:36:04.657123 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xndwx\" (UniqueName: \"kubernetes.io/projected/83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91-kube-api-access-xndwx\") pod \"83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91\" (UID: \"83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91\") " Feb 27 16:36:04 crc kubenswrapper[4814]: I0227 16:36:04.666244 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91-kube-api-access-xndwx" (OuterVolumeSpecName: "kube-api-access-xndwx") pod "83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91" (UID: "83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91"). InnerVolumeSpecName "kube-api-access-xndwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:36:04 crc kubenswrapper[4814]: I0227 16:36:04.759673 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xndwx\" (UniqueName: \"kubernetes.io/projected/83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91-kube-api-access-xndwx\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:05 crc kubenswrapper[4814]: I0227 16:36:05.178072 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536836-7pdzj" event={"ID":"83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91","Type":"ContainerDied","Data":"fe619277ec30081c7b4d088516a652727509d271a3c1d7dd365ba01bf701d716"} Feb 27 16:36:05 crc kubenswrapper[4814]: I0227 16:36:05.178508 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe619277ec30081c7b4d088516a652727509d271a3c1d7dd365ba01bf701d716" Feb 27 16:36:05 crc kubenswrapper[4814]: I0227 16:36:05.178209 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536836-7pdzj" Feb 27 16:36:05 crc kubenswrapper[4814]: I0227 16:36:05.246034 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536830-xr849"] Feb 27 16:36:05 crc kubenswrapper[4814]: I0227 16:36:05.250497 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536830-xr849"] Feb 27 16:36:06 crc kubenswrapper[4814]: I0227 16:36:06.495722 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9c67f63-addd-4849-8c2d-3ea2101d7802" path="/var/lib/kubelet/pods/c9c67f63-addd-4849-8c2d-3ea2101d7802/volumes" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.461952 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-77gdd"] Feb 27 16:36:14 crc kubenswrapper[4814]: E0227 16:36:14.462770 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91" containerName="oc" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.462790 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91" containerName="oc" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.462926 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91" containerName="oc" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.463429 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-77gdd" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.466799 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.467067 4814 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-vldhb" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.470771 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.480269 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-77gdd"] Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.493003 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-wnnxz"] Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.493682 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-wnnxz" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.495355 4814 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-m25dp" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.496935 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-lpq28"] Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.497506 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-lpq28" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.500326 4814 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-ctpsm" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.508126 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-wnnxz"] Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.511622 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-lpq28"] Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.598772 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8pfx\" (UniqueName: \"kubernetes.io/projected/b97362bd-6925-4d89-aba9-a8ad9a9cb1bb-kube-api-access-r8pfx\") pod \"cert-manager-webhook-687f57d79b-lpq28\" (UID: \"b97362bd-6925-4d89-aba9-a8ad9a9cb1bb\") " pod="cert-manager/cert-manager-webhook-687f57d79b-lpq28" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.598881 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p5v5\" (UniqueName: \"kubernetes.io/projected/c6ddbede-41e8-4a88-8348-a81b242a9b85-kube-api-access-8p5v5\") pod \"cert-manager-cainjector-cf98fcc89-77gdd\" (UID: \"c6ddbede-41e8-4a88-8348-a81b242a9b85\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-77gdd" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.598911 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsfw8\" (UniqueName: \"kubernetes.io/projected/9043722e-d6cc-495d-bcf6-ad5b95545ca6-kube-api-access-rsfw8\") pod \"cert-manager-858654f9db-wnnxz\" (UID: \"9043722e-d6cc-495d-bcf6-ad5b95545ca6\") " pod="cert-manager/cert-manager-858654f9db-wnnxz" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.700731 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8pfx\" (UniqueName: \"kubernetes.io/projected/b97362bd-6925-4d89-aba9-a8ad9a9cb1bb-kube-api-access-r8pfx\") pod \"cert-manager-webhook-687f57d79b-lpq28\" (UID: \"b97362bd-6925-4d89-aba9-a8ad9a9cb1bb\") " pod="cert-manager/cert-manager-webhook-687f57d79b-lpq28" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.700871 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p5v5\" (UniqueName: \"kubernetes.io/projected/c6ddbede-41e8-4a88-8348-a81b242a9b85-kube-api-access-8p5v5\") pod \"cert-manager-cainjector-cf98fcc89-77gdd\" (UID: \"c6ddbede-41e8-4a88-8348-a81b242a9b85\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-77gdd" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.700941 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsfw8\" (UniqueName: \"kubernetes.io/projected/9043722e-d6cc-495d-bcf6-ad5b95545ca6-kube-api-access-rsfw8\") pod \"cert-manager-858654f9db-wnnxz\" (UID: \"9043722e-d6cc-495d-bcf6-ad5b95545ca6\") " pod="cert-manager/cert-manager-858654f9db-wnnxz" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.728737 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8pfx\" (UniqueName: \"kubernetes.io/projected/b97362bd-6925-4d89-aba9-a8ad9a9cb1bb-kube-api-access-r8pfx\") pod \"cert-manager-webhook-687f57d79b-lpq28\" (UID: \"b97362bd-6925-4d89-aba9-a8ad9a9cb1bb\") " pod="cert-manager/cert-manager-webhook-687f57d79b-lpq28" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.729174 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsfw8\" (UniqueName: \"kubernetes.io/projected/9043722e-d6cc-495d-bcf6-ad5b95545ca6-kube-api-access-rsfw8\") pod \"cert-manager-858654f9db-wnnxz\" (UID: \"9043722e-d6cc-495d-bcf6-ad5b95545ca6\") " pod="cert-manager/cert-manager-858654f9db-wnnxz" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.731692 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p5v5\" (UniqueName: \"kubernetes.io/projected/c6ddbede-41e8-4a88-8348-a81b242a9b85-kube-api-access-8p5v5\") pod \"cert-manager-cainjector-cf98fcc89-77gdd\" (UID: \"c6ddbede-41e8-4a88-8348-a81b242a9b85\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-77gdd" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.786699 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-77gdd" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.809063 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-wnnxz" Feb 27 16:36:14 crc kubenswrapper[4814]: I0227 16:36:14.815907 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-lpq28" Feb 27 16:36:15 crc kubenswrapper[4814]: I0227 16:36:15.100881 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-77gdd"] Feb 27 16:36:15 crc kubenswrapper[4814]: I0227 16:36:15.252808 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-77gdd" event={"ID":"c6ddbede-41e8-4a88-8348-a81b242a9b85","Type":"ContainerStarted","Data":"67512ebff6267008155740f34a0ced564cbf748e932fa751f5354b40fb5c1671"} Feb 27 16:36:15 crc kubenswrapper[4814]: I0227 16:36:15.267719 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-wnnxz"] Feb 27 16:36:15 crc kubenswrapper[4814]: W0227 16:36:15.277420 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9043722e_d6cc_495d_bcf6_ad5b95545ca6.slice/crio-7020e1733b2045677bf673e9ff70576118a7bfc11974abcd2e9c47f3df7d153f WatchSource:0}: Error finding container 7020e1733b2045677bf673e9ff70576118a7bfc11974abcd2e9c47f3df7d153f: Status 404 returned error can't find the container with id 7020e1733b2045677bf673e9ff70576118a7bfc11974abcd2e9c47f3df7d153f Feb 27 16:36:15 crc kubenswrapper[4814]: I0227 16:36:15.392134 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-lpq28"] Feb 27 16:36:15 crc kubenswrapper[4814]: W0227 16:36:15.394972 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb97362bd_6925_4d89_aba9_a8ad9a9cb1bb.slice/crio-58f0acaf2f97e304a9c3edae442a9a0b4195b587463ad8797b211c270da5cc79 WatchSource:0}: Error finding container 58f0acaf2f97e304a9c3edae442a9a0b4195b587463ad8797b211c270da5cc79: Status 404 returned error can't find the container with id 58f0acaf2f97e304a9c3edae442a9a0b4195b587463ad8797b211c270da5cc79 Feb 27 16:36:16 crc kubenswrapper[4814]: I0227 16:36:16.262046 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-lpq28" event={"ID":"b97362bd-6925-4d89-aba9-a8ad9a9cb1bb","Type":"ContainerStarted","Data":"58f0acaf2f97e304a9c3edae442a9a0b4195b587463ad8797b211c270da5cc79"} Feb 27 16:36:16 crc kubenswrapper[4814]: I0227 16:36:16.264379 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-wnnxz" event={"ID":"9043722e-d6cc-495d-bcf6-ad5b95545ca6","Type":"ContainerStarted","Data":"7020e1733b2045677bf673e9ff70576118a7bfc11974abcd2e9c47f3df7d153f"} Feb 27 16:36:19 crc kubenswrapper[4814]: I0227 16:36:19.293555 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-77gdd" event={"ID":"c6ddbede-41e8-4a88-8348-a81b242a9b85","Type":"ContainerStarted","Data":"14bb43e291b63664f1f843a37a1f81822e44a117ecdc48411be2fb314454f231"} Feb 27 16:36:19 crc kubenswrapper[4814]: I0227 16:36:19.316504 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-77gdd" podStartSLOduration=1.938942824 podStartE2EDuration="5.316485655s" podCreationTimestamp="2026-02-27 16:36:14 +0000 UTC" firstStartedPulling="2026-02-27 16:36:15.112400439 +0000 UTC m=+787.565025269" lastFinishedPulling="2026-02-27 16:36:18.48994326 +0000 UTC m=+790.942568100" observedRunningTime="2026-02-27 16:36:19.312135394 +0000 UTC m=+791.764760274" watchObservedRunningTime="2026-02-27 16:36:19.316485655 +0000 UTC m=+791.769110485" Feb 27 16:36:21 crc kubenswrapper[4814]: I0227 16:36:21.322487 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-wnnxz" event={"ID":"9043722e-d6cc-495d-bcf6-ad5b95545ca6","Type":"ContainerStarted","Data":"8364c2d2429f2aecf34fd2ee0035be6b1a4bbcad816e9100707072d365663a29"} Feb 27 16:36:21 crc kubenswrapper[4814]: I0227 16:36:21.324947 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-lpq28" event={"ID":"b97362bd-6925-4d89-aba9-a8ad9a9cb1bb","Type":"ContainerStarted","Data":"4b098a2c3b9317a394faf318d9d5fe67a286353a9ebad470d456ee368c6ed448"} Feb 27 16:36:21 crc kubenswrapper[4814]: I0227 16:36:21.325189 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-lpq28" Feb 27 16:36:21 crc kubenswrapper[4814]: I0227 16:36:21.346849 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-wnnxz" podStartSLOduration=2.114530987 podStartE2EDuration="7.346830775s" podCreationTimestamp="2026-02-27 16:36:14 +0000 UTC" firstStartedPulling="2026-02-27 16:36:15.281567919 +0000 UTC m=+787.734192779" lastFinishedPulling="2026-02-27 16:36:20.513867697 +0000 UTC m=+792.966492567" observedRunningTime="2026-02-27 16:36:21.345632059 +0000 UTC m=+793.798256919" watchObservedRunningTime="2026-02-27 16:36:21.346830775 +0000 UTC m=+793.799455605" Feb 27 16:36:21 crc kubenswrapper[4814]: I0227 16:36:21.382442 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-lpq28" podStartSLOduration=2.311584275 podStartE2EDuration="7.382417856s" podCreationTimestamp="2026-02-27 16:36:14 +0000 UTC" firstStartedPulling="2026-02-27 16:36:15.398868917 +0000 UTC m=+787.851493757" lastFinishedPulling="2026-02-27 16:36:20.469702478 +0000 UTC m=+792.922327338" observedRunningTime="2026-02-27 16:36:21.373718494 +0000 UTC m=+793.826343364" watchObservedRunningTime="2026-02-27 16:36:21.382417856 +0000 UTC m=+793.835042726" Feb 27 16:36:23 crc kubenswrapper[4814]: I0227 16:36:23.635929 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7zrr2"] Feb 27 16:36:23 crc kubenswrapper[4814]: I0227 16:36:23.637569 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovn-controller" containerID="cri-o://d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020" gracePeriod=30 Feb 27 16:36:23 crc kubenswrapper[4814]: I0227 16:36:23.637620 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e" gracePeriod=30 Feb 27 16:36:23 crc kubenswrapper[4814]: I0227 16:36:23.637658 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="sbdb" containerID="cri-o://936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1" gracePeriod=30 Feb 27 16:36:23 crc kubenswrapper[4814]: I0227 16:36:23.637583 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="nbdb" containerID="cri-o://4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66" gracePeriod=30 Feb 27 16:36:23 crc kubenswrapper[4814]: I0227 16:36:23.637744 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovn-acl-logging" containerID="cri-o://aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825" gracePeriod=30 Feb 27 16:36:23 crc kubenswrapper[4814]: I0227 16:36:23.637780 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="northd" containerID="cri-o://9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1" gracePeriod=30 Feb 27 16:36:23 crc kubenswrapper[4814]: I0227 16:36:23.637789 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="kube-rbac-proxy-node" containerID="cri-o://b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2" gracePeriod=30 Feb 27 16:36:23 crc kubenswrapper[4814]: I0227 16:36:23.698897 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovnkube-controller" containerID="cri-o://bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a" gracePeriod=30 Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.015422 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovnkube-controller/3.log" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.018213 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovn-acl-logging/0.log" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.018820 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovn-controller/0.log" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.019422 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040552 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-ovnkube-config\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040618 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-ovnkube-script-lib\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040642 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-systemd-units\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040669 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-cni-bin\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040709 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jntcv\" (UniqueName: \"kubernetes.io/projected/4619e86c-7eae-4849-89b0-abb44778d871-kube-api-access-jntcv\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040741 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-var-lib-openvswitch\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040759 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-etc-openvswitch\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040777 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-cni-netd\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040797 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-var-lib-cni-networks-ovn-kubernetes\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040876 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040875 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040919 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040928 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040947 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040960 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040999 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.040977 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-openvswitch\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041065 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-log-socket\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041151 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4619e86c-7eae-4849-89b0-abb44778d871-ovn-node-metrics-cert\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041191 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-kubelet\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041211 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-ovn\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041212 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-log-socket" (OuterVolumeSpecName: "log-socket") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041239 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-env-overrides\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041298 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-run-netns\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041293 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041310 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041356 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-run-ovn-kubernetes\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041380 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041405 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-systemd\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041428 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-slash\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041459 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-node-log\") pod \"4619e86c-7eae-4849-89b0-abb44778d871\" (UID: \"4619e86c-7eae-4849-89b0-abb44778d871\") " Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041511 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-slash" (OuterVolumeSpecName: "host-slash") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041758 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041831 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-node-log" (OuterVolumeSpecName: "node-log") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041892 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041943 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.042121 4814 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.042142 4814 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.042153 4814 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.042166 4814 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.042180 4814 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.042192 4814 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-log-socket\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.042204 4814 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.042215 4814 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.042226 4814 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.042238 4814 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.042249 4814 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-slash\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.042277 4814 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-node-log\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.042288 4814 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.042299 4814 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4619e86c-7eae-4849-89b0-abb44778d871-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.042310 4814 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.042320 4814 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.041430 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.047469 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4619e86c-7eae-4849-89b0-abb44778d871-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.047699 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4619e86c-7eae-4849-89b0-abb44778d871-kube-api-access-jntcv" (OuterVolumeSpecName: "kube-api-access-jntcv") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "kube-api-access-jntcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.060473 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "4619e86c-7eae-4849-89b0-abb44778d871" (UID: "4619e86c-7eae-4849-89b0-abb44778d871"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.090704 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-54hbp"] Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.091025 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovnkube-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091055 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovnkube-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.091075 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovn-acl-logging" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091089 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovn-acl-logging" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.091105 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovnkube-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091122 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovnkube-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.091142 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="nbdb" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091157 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="nbdb" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.091174 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovnkube-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091187 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovnkube-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.091205 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovnkube-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091218 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovnkube-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.091236 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="kube-rbac-proxy-ovn-metrics" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091249 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="kube-rbac-proxy-ovn-metrics" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.091295 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="kube-rbac-proxy-node" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091309 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="kube-rbac-proxy-node" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.091328 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovn-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091341 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovn-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.091369 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="northd" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091382 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="northd" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.091399 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="kubecfg-setup" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091412 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="kubecfg-setup" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.091438 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="sbdb" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091452 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="sbdb" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091624 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovnkube-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091642 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovnkube-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091660 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovn-acl-logging" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091675 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="kube-rbac-proxy-node" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091695 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovn-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091710 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovnkube-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091729 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="northd" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091745 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="nbdb" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091763 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovnkube-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091778 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovnkube-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091796 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="kube-rbac-proxy-ovn-metrics" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.091816 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="sbdb" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.091985 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovnkube-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.092000 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4619e86c-7eae-4849-89b0-abb44778d871" containerName="ovnkube-controller" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.095471 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144014 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-var-lib-openvswitch\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144131 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-slash\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144181 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-log-socket\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144233 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-etc-openvswitch\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144300 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-node-log\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144333 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144372 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-run-systemd\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144481 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-kubelet\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144575 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-run-openvswitch\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144641 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9efc887f-16f7-46b1-bf6d-748a069e00c5-env-overrides\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144672 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-cni-netd\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144774 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-systemd-units\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144836 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9efc887f-16f7-46b1-bf6d-748a069e00c5-ovn-node-metrics-cert\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144869 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9efc887f-16f7-46b1-bf6d-748a069e00c5-ovnkube-script-lib\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144922 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-run-ovn-kubernetes\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144958 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9efc887f-16f7-46b1-bf6d-748a069e00c5-ovnkube-config\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.144994 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr6c5\" (UniqueName: \"kubernetes.io/projected/9efc887f-16f7-46b1-bf6d-748a069e00c5-kube-api-access-pr6c5\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.145128 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-run-netns\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.145164 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-run-ovn\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.145210 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-cni-bin\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.145308 4814 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.145334 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jntcv\" (UniqueName: \"kubernetes.io/projected/4619e86c-7eae-4849-89b0-abb44778d871-kube-api-access-jntcv\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.145357 4814 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4619e86c-7eae-4849-89b0-abb44778d871-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.145378 4814 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4619e86c-7eae-4849-89b0-abb44778d871-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.245685 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-log-socket\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.245742 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-etc-openvswitch\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.245767 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.245789 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-node-log\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.245811 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-run-systemd\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.245842 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-kubelet\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.245863 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-run-openvswitch\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.245885 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9efc887f-16f7-46b1-bf6d-748a069e00c5-env-overrides\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.245904 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-cni-netd\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.245922 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-systemd-units\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.245943 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9efc887f-16f7-46b1-bf6d-748a069e00c5-ovn-node-metrics-cert\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.245942 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-etc-openvswitch\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.245984 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-kubelet\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.245994 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-log-socket\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246049 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-cni-netd\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246075 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-systemd-units\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246081 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-run-systemd\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246013 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-node-log\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246058 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246107 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-run-openvswitch\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.245965 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9efc887f-16f7-46b1-bf6d-748a069e00c5-ovnkube-script-lib\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246310 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-run-ovn-kubernetes\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246348 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9efc887f-16f7-46b1-bf6d-748a069e00c5-ovnkube-config\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246370 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-run-ovn-kubernetes\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246393 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr6c5\" (UniqueName: \"kubernetes.io/projected/9efc887f-16f7-46b1-bf6d-748a069e00c5-kube-api-access-pr6c5\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246520 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-run-ovn\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246556 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-run-netns\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246588 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-cni-bin\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246629 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-var-lib-openvswitch\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246662 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-slash\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246778 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-slash\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246802 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9efc887f-16f7-46b1-bf6d-748a069e00c5-ovnkube-script-lib\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246852 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-run-netns\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246885 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-run-ovn\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246912 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-host-cni-bin\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.246938 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9efc887f-16f7-46b1-bf6d-748a069e00c5-var-lib-openvswitch\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.247076 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9efc887f-16f7-46b1-bf6d-748a069e00c5-ovnkube-config\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.247172 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9efc887f-16f7-46b1-bf6d-748a069e00c5-env-overrides\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.250066 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9efc887f-16f7-46b1-bf6d-748a069e00c5-ovn-node-metrics-cert\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.275220 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr6c5\" (UniqueName: \"kubernetes.io/projected/9efc887f-16f7-46b1-bf6d-748a069e00c5-kube-api-access-pr6c5\") pod \"ovnkube-node-54hbp\" (UID: \"9efc887f-16f7-46b1-bf6d-748a069e00c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.356340 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovnkube-controller/3.log" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.366721 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovn-acl-logging/0.log" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.368645 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zrr2_4619e86c-7eae-4849-89b0-abb44778d871/ovn-controller/0.log" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373018 4814 generic.go:334] "Generic (PLEG): container finished" podID="4619e86c-7eae-4849-89b0-abb44778d871" containerID="bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a" exitCode=0 Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373097 4814 generic.go:334] "Generic (PLEG): container finished" podID="4619e86c-7eae-4849-89b0-abb44778d871" containerID="936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1" exitCode=0 Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373124 4814 generic.go:334] "Generic (PLEG): container finished" podID="4619e86c-7eae-4849-89b0-abb44778d871" containerID="4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66" exitCode=0 Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373142 4814 generic.go:334] "Generic (PLEG): container finished" podID="4619e86c-7eae-4849-89b0-abb44778d871" containerID="9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1" exitCode=0 Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373125 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerDied","Data":"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373223 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerDied","Data":"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373286 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerDied","Data":"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373166 4814 generic.go:334] "Generic (PLEG): container finished" podID="4619e86c-7eae-4849-89b0-abb44778d871" containerID="aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e" exitCode=0 Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373338 4814 generic.go:334] "Generic (PLEG): container finished" podID="4619e86c-7eae-4849-89b0-abb44778d871" containerID="b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2" exitCode=0 Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373362 4814 generic.go:334] "Generic (PLEG): container finished" podID="4619e86c-7eae-4849-89b0-abb44778d871" containerID="aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825" exitCode=143 Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373383 4814 generic.go:334] "Generic (PLEG): container finished" podID="4619e86c-7eae-4849-89b0-abb44778d871" containerID="d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020" exitCode=143 Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373317 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerDied","Data":"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373339 4814 scope.go:117] "RemoveContainer" containerID="bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373503 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerDied","Data":"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373547 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerDied","Data":"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373606 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373625 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373778 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373793 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373804 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373815 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373861 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373874 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373885 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373906 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerDied","Data":"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373960 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373977 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.373989 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.374000 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.374069 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.374191 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.374205 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.374244 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.374302 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.374315 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.374394 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.374332 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerDied","Data":"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.374938 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.374967 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.374979 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.374993 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375004 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375015 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375028 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375039 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375050 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375093 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375210 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zrr2" event={"ID":"4619e86c-7eae-4849-89b0-abb44778d871","Type":"ContainerDied","Data":"bd2b250dddc2ff4cf4bea73006bea42adf0ef0909bde6e8efef659d20f20f032"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375294 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375313 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375325 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375343 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375396 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375409 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375422 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375434 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375478 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.375491 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.379042 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4m267_56739684-228b-4bf3-bb4e-11f7aaca1890/kube-multus/2.log" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.379972 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4m267_56739684-228b-4bf3-bb4e-11f7aaca1890/kube-multus/1.log" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.380108 4814 generic.go:334] "Generic (PLEG): container finished" podID="56739684-228b-4bf3-bb4e-11f7aaca1890" containerID="d7cc356e7ca5938965550870658530bc06e9d97a37f7b465b73ef82649dcd82e" exitCode=2 Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.380183 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4m267" event={"ID":"56739684-228b-4bf3-bb4e-11f7aaca1890","Type":"ContainerDied","Data":"d7cc356e7ca5938965550870658530bc06e9d97a37f7b465b73ef82649dcd82e"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.380223 4814 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c7417cf941b0d6c4cba975b6596d39476bda8e19c0399f28a0288e09bc436549"} Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.381113 4814 scope.go:117] "RemoveContainer" containerID="d7cc356e7ca5938965550870658530bc06e9d97a37f7b465b73ef82649dcd82e" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.381569 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-4m267_openshift-multus(56739684-228b-4bf3-bb4e-11f7aaca1890)\"" pod="openshift-multus/multus-4m267" podUID="56739684-228b-4bf3-bb4e-11f7aaca1890" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.398246 4814 scope.go:117] "RemoveContainer" containerID="17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.417937 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.437569 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7zrr2"] Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.442315 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7zrr2"] Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.468653 4814 scope.go:117] "RemoveContainer" containerID="936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.489020 4814 scope.go:117] "RemoveContainer" containerID="4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.497599 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4619e86c-7eae-4849-89b0-abb44778d871" path="/var/lib/kubelet/pods/4619e86c-7eae-4849-89b0-abb44778d871/volumes" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.508000 4814 scope.go:117] "RemoveContainer" containerID="9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.532154 4814 scope.go:117] "RemoveContainer" containerID="aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.558691 4814 scope.go:117] "RemoveContainer" containerID="b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.581184 4814 scope.go:117] "RemoveContainer" containerID="aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.600079 4814 scope.go:117] "RemoveContainer" containerID="d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.630582 4814 scope.go:117] "RemoveContainer" containerID="5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.699425 4814 scope.go:117] "RemoveContainer" containerID="bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.699902 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a\": container with ID starting with bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a not found: ID does not exist" containerID="bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.699949 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a"} err="failed to get container status \"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a\": rpc error: code = NotFound desc = could not find container \"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a\": container with ID starting with bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.699981 4814 scope.go:117] "RemoveContainer" containerID="17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.700495 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd\": container with ID starting with 17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd not found: ID does not exist" containerID="17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.700527 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd"} err="failed to get container status \"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd\": rpc error: code = NotFound desc = could not find container \"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd\": container with ID starting with 17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.700551 4814 scope.go:117] "RemoveContainer" containerID="936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.700859 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\": container with ID starting with 936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1 not found: ID does not exist" containerID="936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.700907 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1"} err="failed to get container status \"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\": rpc error: code = NotFound desc = could not find container \"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\": container with ID starting with 936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.700922 4814 scope.go:117] "RemoveContainer" containerID="4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.701456 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\": container with ID starting with 4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66 not found: ID does not exist" containerID="4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.701482 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66"} err="failed to get container status \"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\": rpc error: code = NotFound desc = could not find container \"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\": container with ID starting with 4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.701498 4814 scope.go:117] "RemoveContainer" containerID="9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.701866 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\": container with ID starting with 9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1 not found: ID does not exist" containerID="9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.701896 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1"} err="failed to get container status \"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\": rpc error: code = NotFound desc = could not find container \"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\": container with ID starting with 9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.701911 4814 scope.go:117] "RemoveContainer" containerID="aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.702186 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\": container with ID starting with aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e not found: ID does not exist" containerID="aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.702241 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e"} err="failed to get container status \"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\": rpc error: code = NotFound desc = could not find container \"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\": container with ID starting with aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.702311 4814 scope.go:117] "RemoveContainer" containerID="b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.702637 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\": container with ID starting with b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2 not found: ID does not exist" containerID="b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.702660 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2"} err="failed to get container status \"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\": rpc error: code = NotFound desc = could not find container \"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\": container with ID starting with b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.702677 4814 scope.go:117] "RemoveContainer" containerID="aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.702958 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\": container with ID starting with aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825 not found: ID does not exist" containerID="aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.702992 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825"} err="failed to get container status \"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\": rpc error: code = NotFound desc = could not find container \"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\": container with ID starting with aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.703016 4814 scope.go:117] "RemoveContainer" containerID="d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.703304 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\": container with ID starting with d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020 not found: ID does not exist" containerID="d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.703332 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020"} err="failed to get container status \"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\": rpc error: code = NotFound desc = could not find container \"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\": container with ID starting with d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.703352 4814 scope.go:117] "RemoveContainer" containerID="5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42" Feb 27 16:36:24 crc kubenswrapper[4814]: E0227 16:36:24.703591 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\": container with ID starting with 5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42 not found: ID does not exist" containerID="5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.703628 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42"} err="failed to get container status \"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\": rpc error: code = NotFound desc = could not find container \"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\": container with ID starting with 5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.703655 4814 scope.go:117] "RemoveContainer" containerID="bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.703905 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a"} err="failed to get container status \"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a\": rpc error: code = NotFound desc = could not find container \"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a\": container with ID starting with bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.703958 4814 scope.go:117] "RemoveContainer" containerID="17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.704329 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd"} err="failed to get container status \"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd\": rpc error: code = NotFound desc = could not find container \"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd\": container with ID starting with 17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.704346 4814 scope.go:117] "RemoveContainer" containerID="936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.704592 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1"} err="failed to get container status \"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\": rpc error: code = NotFound desc = could not find container \"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\": container with ID starting with 936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.704606 4814 scope.go:117] "RemoveContainer" containerID="4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.704955 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66"} err="failed to get container status \"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\": rpc error: code = NotFound desc = could not find container \"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\": container with ID starting with 4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.704969 4814 scope.go:117] "RemoveContainer" containerID="9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.705204 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1"} err="failed to get container status \"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\": rpc error: code = NotFound desc = could not find container \"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\": container with ID starting with 9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.705446 4814 scope.go:117] "RemoveContainer" containerID="aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.705723 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e"} err="failed to get container status \"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\": rpc error: code = NotFound desc = could not find container \"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\": container with ID starting with aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.705777 4814 scope.go:117] "RemoveContainer" containerID="b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.706097 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2"} err="failed to get container status \"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\": rpc error: code = NotFound desc = could not find container \"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\": container with ID starting with b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.706123 4814 scope.go:117] "RemoveContainer" containerID="aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.706339 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825"} err="failed to get container status \"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\": rpc error: code = NotFound desc = could not find container \"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\": container with ID starting with aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.706353 4814 scope.go:117] "RemoveContainer" containerID="d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.706543 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020"} err="failed to get container status \"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\": rpc error: code = NotFound desc = could not find container \"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\": container with ID starting with d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.706558 4814 scope.go:117] "RemoveContainer" containerID="5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.706809 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42"} err="failed to get container status \"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\": rpc error: code = NotFound desc = could not find container \"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\": container with ID starting with 5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.706824 4814 scope.go:117] "RemoveContainer" containerID="bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.707147 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a"} err="failed to get container status \"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a\": rpc error: code = NotFound desc = could not find container \"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a\": container with ID starting with bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.707168 4814 scope.go:117] "RemoveContainer" containerID="17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.707407 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd"} err="failed to get container status \"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd\": rpc error: code = NotFound desc = could not find container \"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd\": container with ID starting with 17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.707431 4814 scope.go:117] "RemoveContainer" containerID="936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.707696 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1"} err="failed to get container status \"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\": rpc error: code = NotFound desc = could not find container \"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\": container with ID starting with 936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.707713 4814 scope.go:117] "RemoveContainer" containerID="4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.707921 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66"} err="failed to get container status \"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\": rpc error: code = NotFound desc = could not find container \"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\": container with ID starting with 4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.707934 4814 scope.go:117] "RemoveContainer" containerID="9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.708152 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1"} err="failed to get container status \"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\": rpc error: code = NotFound desc = could not find container \"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\": container with ID starting with 9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.708167 4814 scope.go:117] "RemoveContainer" containerID="aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.708359 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e"} err="failed to get container status \"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\": rpc error: code = NotFound desc = could not find container \"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\": container with ID starting with aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.708455 4814 scope.go:117] "RemoveContainer" containerID="b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.708649 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2"} err="failed to get container status \"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\": rpc error: code = NotFound desc = could not find container \"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\": container with ID starting with b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.708677 4814 scope.go:117] "RemoveContainer" containerID="aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.708886 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825"} err="failed to get container status \"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\": rpc error: code = NotFound desc = could not find container \"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\": container with ID starting with aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.708932 4814 scope.go:117] "RemoveContainer" containerID="d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.709189 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020"} err="failed to get container status \"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\": rpc error: code = NotFound desc = could not find container \"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\": container with ID starting with d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.709220 4814 scope.go:117] "RemoveContainer" containerID="5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.709450 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42"} err="failed to get container status \"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\": rpc error: code = NotFound desc = could not find container \"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\": container with ID starting with 5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.709469 4814 scope.go:117] "RemoveContainer" containerID="bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.710065 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a"} err="failed to get container status \"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a\": rpc error: code = NotFound desc = could not find container \"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a\": container with ID starting with bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.710111 4814 scope.go:117] "RemoveContainer" containerID="17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.710428 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd"} err="failed to get container status \"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd\": rpc error: code = NotFound desc = could not find container \"17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd\": container with ID starting with 17ecd712da49e2ba02a311fa2bcb29a73ef89561ed56bbcdbd8db983c08f5ddd not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.710454 4814 scope.go:117] "RemoveContainer" containerID="936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.710646 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1"} err="failed to get container status \"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\": rpc error: code = NotFound desc = could not find container \"936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1\": container with ID starting with 936921f18cb9c00ae2a45240c336a04dd414b5fc45bbee4748b8cf8f9b9541e1 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.710677 4814 scope.go:117] "RemoveContainer" containerID="4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.710906 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66"} err="failed to get container status \"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\": rpc error: code = NotFound desc = could not find container \"4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66\": container with ID starting with 4112cacfb490e36b54d79010c9d5bfe5dc7f57398fffb66c0465e11f99873b66 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.710936 4814 scope.go:117] "RemoveContainer" containerID="9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.711220 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1"} err="failed to get container status \"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\": rpc error: code = NotFound desc = could not find container \"9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1\": container with ID starting with 9454a4f4276683503b10dd3512e5dd700ffaf0741642425e9e8ca9e2e693bad1 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.711241 4814 scope.go:117] "RemoveContainer" containerID="aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.711438 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e"} err="failed to get container status \"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\": rpc error: code = NotFound desc = could not find container \"aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e\": container with ID starting with aad82e44e9c222b4ee6a2eabb9d90127fd323ebaf2562a33c70fb8d3b800f58e not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.711457 4814 scope.go:117] "RemoveContainer" containerID="b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.711640 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2"} err="failed to get container status \"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\": rpc error: code = NotFound desc = could not find container \"b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2\": container with ID starting with b9222198eeaec32809e2d6d53be539105fd96e42724e8dd901e0399ee55fd6c2 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.711658 4814 scope.go:117] "RemoveContainer" containerID="aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.711932 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825"} err="failed to get container status \"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\": rpc error: code = NotFound desc = could not find container \"aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825\": container with ID starting with aa79171f8208bf10507f44568ae97aaba50ac6686bce2f58f141c8cdbbbcc825 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.711955 4814 scope.go:117] "RemoveContainer" containerID="d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.712171 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020"} err="failed to get container status \"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\": rpc error: code = NotFound desc = could not find container \"d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020\": container with ID starting with d857ff89c994244592e954fe89b733449f9c2aa0a36ded606a20b35fd735b020 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.712185 4814 scope.go:117] "RemoveContainer" containerID="5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.712444 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42"} err="failed to get container status \"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\": rpc error: code = NotFound desc = could not find container \"5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42\": container with ID starting with 5dc0305eb74142ee2e7e57399f411238b9d32cbdead50aa298443df106d86a42 not found: ID does not exist" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.712464 4814 scope.go:117] "RemoveContainer" containerID="bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a" Feb 27 16:36:24 crc kubenswrapper[4814]: I0227 16:36:24.712773 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a"} err="failed to get container status \"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a\": rpc error: code = NotFound desc = could not find container \"bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a\": container with ID starting with bdfd56e456e0ad044c7165841c21cc7986ddf53bba102eac805d26c2b9d4f98a not found: ID does not exist" Feb 27 16:36:25 crc kubenswrapper[4814]: I0227 16:36:25.388962 4814 generic.go:334] "Generic (PLEG): container finished" podID="9efc887f-16f7-46b1-bf6d-748a069e00c5" containerID="7f723a45acbc5f1d95276848b01e78d745e9b7adf67886794a2df4a0e279bd95" exitCode=0 Feb 27 16:36:25 crc kubenswrapper[4814]: I0227 16:36:25.389099 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" event={"ID":"9efc887f-16f7-46b1-bf6d-748a069e00c5","Type":"ContainerDied","Data":"7f723a45acbc5f1d95276848b01e78d745e9b7adf67886794a2df4a0e279bd95"} Feb 27 16:36:25 crc kubenswrapper[4814]: I0227 16:36:25.389668 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" event={"ID":"9efc887f-16f7-46b1-bf6d-748a069e00c5","Type":"ContainerStarted","Data":"f433f1e2588e09b79f109a5c5bd27b8fcb43039814e000a81605210d9ba1048e"} Feb 27 16:36:26 crc kubenswrapper[4814]: I0227 16:36:26.012188 4814 scope.go:117] "RemoveContainer" containerID="74afe3da5ba4c9243941b03926af4ac2b8cf13e618d045c092c82ffab6cc3ae7" Feb 27 16:36:26 crc kubenswrapper[4814]: I0227 16:36:26.035687 4814 scope.go:117] "RemoveContainer" containerID="f54787b0135ae6183bcf18cdd7f2beef6bd7b6d884e1079dd611efbdfe04e5c8" Feb 27 16:36:26 crc kubenswrapper[4814]: I0227 16:36:26.090351 4814 scope.go:117] "RemoveContainer" containerID="c7417cf941b0d6c4cba975b6596d39476bda8e19c0399f28a0288e09bc436549" Feb 27 16:36:26 crc kubenswrapper[4814]: I0227 16:36:26.410630 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" event={"ID":"9efc887f-16f7-46b1-bf6d-748a069e00c5","Type":"ContainerStarted","Data":"61458dba4d80ba216d46d83a149358e94ed1cbeccc76e44002e88d4ca2e29c7e"} Feb 27 16:36:26 crc kubenswrapper[4814]: I0227 16:36:26.411118 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" event={"ID":"9efc887f-16f7-46b1-bf6d-748a069e00c5","Type":"ContainerStarted","Data":"cdc6d90a8ec94ccdf01d666595bf451bc7a2a9f7a6b3fff4de7a00ac73c0726b"} Feb 27 16:36:26 crc kubenswrapper[4814]: I0227 16:36:26.411143 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" event={"ID":"9efc887f-16f7-46b1-bf6d-748a069e00c5","Type":"ContainerStarted","Data":"40ec76a9b2ee64d2979eb093bede5dd8b9279b08e34879db43bab699c7cebd3f"} Feb 27 16:36:26 crc kubenswrapper[4814]: I0227 16:36:26.411162 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" event={"ID":"9efc887f-16f7-46b1-bf6d-748a069e00c5","Type":"ContainerStarted","Data":"80f0eb78ac6b865538697a153a08aef746220d6c208236a889127e926d7f5598"} Feb 27 16:36:26 crc kubenswrapper[4814]: I0227 16:36:26.411187 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" event={"ID":"9efc887f-16f7-46b1-bf6d-748a069e00c5","Type":"ContainerStarted","Data":"7c1b7f89ef22f56bd4b249c050d50ace552e955d749f2960f8b0eae380414721"} Feb 27 16:36:26 crc kubenswrapper[4814]: I0227 16:36:26.411204 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" event={"ID":"9efc887f-16f7-46b1-bf6d-748a069e00c5","Type":"ContainerStarted","Data":"06ef932a9a7b28ecb6a9d5bb87e8fd8c8af49248536ca864b0363379ab01f48a"} Feb 27 16:36:26 crc kubenswrapper[4814]: I0227 16:36:26.413957 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4m267_56739684-228b-4bf3-bb4e-11f7aaca1890/kube-multus/2.log" Feb 27 16:36:29 crc kubenswrapper[4814]: I0227 16:36:29.464812 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" event={"ID":"9efc887f-16f7-46b1-bf6d-748a069e00c5","Type":"ContainerStarted","Data":"8c45a959cd27ff42e10c5f1a00f4a50bc67ddf7a1e0a6abebc136414140140db"} Feb 27 16:36:29 crc kubenswrapper[4814]: I0227 16:36:29.821447 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-lpq28" Feb 27 16:36:31 crc kubenswrapper[4814]: I0227 16:36:31.489049 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" event={"ID":"9efc887f-16f7-46b1-bf6d-748a069e00c5","Type":"ContainerStarted","Data":"51d2803f20749c6016c086a3b7add13ce06945548dec9270fe253846973d53b7"} Feb 27 16:36:31 crc kubenswrapper[4814]: I0227 16:36:31.489500 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:31 crc kubenswrapper[4814]: I0227 16:36:31.489647 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:31 crc kubenswrapper[4814]: I0227 16:36:31.523059 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" podStartSLOduration=7.52304327 podStartE2EDuration="7.52304327s" podCreationTimestamp="2026-02-27 16:36:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:36:31.522305677 +0000 UTC m=+803.974930517" watchObservedRunningTime="2026-02-27 16:36:31.52304327 +0000 UTC m=+803.975668100" Feb 27 16:36:31 crc kubenswrapper[4814]: I0227 16:36:31.523269 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:32 crc kubenswrapper[4814]: I0227 16:36:32.499998 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:32 crc kubenswrapper[4814]: I0227 16:36:32.547788 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:36:35 crc kubenswrapper[4814]: I0227 16:36:35.487326 4814 scope.go:117] "RemoveContainer" containerID="d7cc356e7ca5938965550870658530bc06e9d97a37f7b465b73ef82649dcd82e" Feb 27 16:36:35 crc kubenswrapper[4814]: E0227 16:36:35.487798 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-4m267_openshift-multus(56739684-228b-4bf3-bb4e-11f7aaca1890)\"" pod="openshift-multus/multus-4m267" podUID="56739684-228b-4bf3-bb4e-11f7aaca1890" Feb 27 16:36:48 crc kubenswrapper[4814]: I0227 16:36:48.494664 4814 scope.go:117] "RemoveContainer" containerID="d7cc356e7ca5938965550870658530bc06e9d97a37f7b465b73ef82649dcd82e" Feb 27 16:36:49 crc kubenswrapper[4814]: I0227 16:36:49.631366 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4m267_56739684-228b-4bf3-bb4e-11f7aaca1890/kube-multus/2.log" Feb 27 16:36:49 crc kubenswrapper[4814]: I0227 16:36:49.631806 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4m267" event={"ID":"56739684-228b-4bf3-bb4e-11f7aaca1890","Type":"ContainerStarted","Data":"d2e1124b8e5e0933babf3c52f577ead222af9bbf83025bbc122af901600a8cbd"} Feb 27 16:36:54 crc kubenswrapper[4814]: I0227 16:36:54.455926 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-54hbp" Feb 27 16:37:05 crc kubenswrapper[4814]: I0227 16:37:05.653967 4814 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 27 16:37:14 crc kubenswrapper[4814]: I0227 16:37:14.725034 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr"] Feb 27 16:37:14 crc kubenswrapper[4814]: I0227 16:37:14.728911 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" Feb 27 16:37:14 crc kubenswrapper[4814]: I0227 16:37:14.732118 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 27 16:37:14 crc kubenswrapper[4814]: I0227 16:37:14.756365 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr"] Feb 27 16:37:14 crc kubenswrapper[4814]: I0227 16:37:14.798782 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4fb39e04-954d-4a82-9d19-ece0f11c5118-bundle\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr\" (UID: \"4fb39e04-954d-4a82-9d19-ece0f11c5118\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" Feb 27 16:37:14 crc kubenswrapper[4814]: I0227 16:37:14.798861 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2hm4\" (UniqueName: \"kubernetes.io/projected/4fb39e04-954d-4a82-9d19-ece0f11c5118-kube-api-access-x2hm4\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr\" (UID: \"4fb39e04-954d-4a82-9d19-ece0f11c5118\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" Feb 27 16:37:14 crc kubenswrapper[4814]: I0227 16:37:14.798970 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4fb39e04-954d-4a82-9d19-ece0f11c5118-util\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr\" (UID: \"4fb39e04-954d-4a82-9d19-ece0f11c5118\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" Feb 27 16:37:14 crc kubenswrapper[4814]: I0227 16:37:14.900664 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4fb39e04-954d-4a82-9d19-ece0f11c5118-bundle\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr\" (UID: \"4fb39e04-954d-4a82-9d19-ece0f11c5118\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" Feb 27 16:37:14 crc kubenswrapper[4814]: I0227 16:37:14.901285 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2hm4\" (UniqueName: \"kubernetes.io/projected/4fb39e04-954d-4a82-9d19-ece0f11c5118-kube-api-access-x2hm4\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr\" (UID: \"4fb39e04-954d-4a82-9d19-ece0f11c5118\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" Feb 27 16:37:14 crc kubenswrapper[4814]: I0227 16:37:14.901564 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4fb39e04-954d-4a82-9d19-ece0f11c5118-util\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr\" (UID: \"4fb39e04-954d-4a82-9d19-ece0f11c5118\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" Feb 27 16:37:14 crc kubenswrapper[4814]: I0227 16:37:14.901577 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4fb39e04-954d-4a82-9d19-ece0f11c5118-bundle\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr\" (UID: \"4fb39e04-954d-4a82-9d19-ece0f11c5118\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" Feb 27 16:37:14 crc kubenswrapper[4814]: I0227 16:37:14.902154 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4fb39e04-954d-4a82-9d19-ece0f11c5118-util\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr\" (UID: \"4fb39e04-954d-4a82-9d19-ece0f11c5118\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" Feb 27 16:37:14 crc kubenswrapper[4814]: I0227 16:37:14.929041 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2hm4\" (UniqueName: \"kubernetes.io/projected/4fb39e04-954d-4a82-9d19-ece0f11c5118-kube-api-access-x2hm4\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr\" (UID: \"4fb39e04-954d-4a82-9d19-ece0f11c5118\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" Feb 27 16:37:15 crc kubenswrapper[4814]: I0227 16:37:15.051945 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" Feb 27 16:37:15 crc kubenswrapper[4814]: I0227 16:37:15.335976 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr"] Feb 27 16:37:15 crc kubenswrapper[4814]: W0227 16:37:15.349473 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4fb39e04_954d_4a82_9d19_ece0f11c5118.slice/crio-a5376bc0efae9864a913d2b5136864a0c1e3e16e1b7a65f411edfa9eddda1784 WatchSource:0}: Error finding container a5376bc0efae9864a913d2b5136864a0c1e3e16e1b7a65f411edfa9eddda1784: Status 404 returned error can't find the container with id a5376bc0efae9864a913d2b5136864a0c1e3e16e1b7a65f411edfa9eddda1784 Feb 27 16:37:15 crc kubenswrapper[4814]: I0227 16:37:15.848720 4814 generic.go:334] "Generic (PLEG): container finished" podID="4fb39e04-954d-4a82-9d19-ece0f11c5118" containerID="1b450fd4b1a3ae3d0488f68bac8071879eea1bc823eb2c65ffc31aebc47ba26f" exitCode=0 Feb 27 16:37:15 crc kubenswrapper[4814]: I0227 16:37:15.848783 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" event={"ID":"4fb39e04-954d-4a82-9d19-ece0f11c5118","Type":"ContainerDied","Data":"1b450fd4b1a3ae3d0488f68bac8071879eea1bc823eb2c65ffc31aebc47ba26f"} Feb 27 16:37:15 crc kubenswrapper[4814]: I0227 16:37:15.848822 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" event={"ID":"4fb39e04-954d-4a82-9d19-ece0f11c5118","Type":"ContainerStarted","Data":"a5376bc0efae9864a913d2b5136864a0c1e3e16e1b7a65f411edfa9eddda1784"} Feb 27 16:37:15 crc kubenswrapper[4814]: I0227 16:37:15.851009 4814 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 16:37:16 crc kubenswrapper[4814]: I0227 16:37:16.174790 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6nmxg"] Feb 27 16:37:16 crc kubenswrapper[4814]: I0227 16:37:16.177614 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:16 crc kubenswrapper[4814]: I0227 16:37:16.189727 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6nmxg"] Feb 27 16:37:16 crc kubenswrapper[4814]: I0227 16:37:16.226494 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2zc4\" (UniqueName: \"kubernetes.io/projected/279f32be-baa9-4cf7-a540-f6f8552c2096-kube-api-access-s2zc4\") pod \"redhat-operators-6nmxg\" (UID: \"279f32be-baa9-4cf7-a540-f6f8552c2096\") " pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:16 crc kubenswrapper[4814]: I0227 16:37:16.226599 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/279f32be-baa9-4cf7-a540-f6f8552c2096-utilities\") pod \"redhat-operators-6nmxg\" (UID: \"279f32be-baa9-4cf7-a540-f6f8552c2096\") " pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:16 crc kubenswrapper[4814]: I0227 16:37:16.226702 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/279f32be-baa9-4cf7-a540-f6f8552c2096-catalog-content\") pod \"redhat-operators-6nmxg\" (UID: \"279f32be-baa9-4cf7-a540-f6f8552c2096\") " pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:16 crc kubenswrapper[4814]: I0227 16:37:16.327880 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2zc4\" (UniqueName: \"kubernetes.io/projected/279f32be-baa9-4cf7-a540-f6f8552c2096-kube-api-access-s2zc4\") pod \"redhat-operators-6nmxg\" (UID: \"279f32be-baa9-4cf7-a540-f6f8552c2096\") " pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:16 crc kubenswrapper[4814]: I0227 16:37:16.328419 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/279f32be-baa9-4cf7-a540-f6f8552c2096-utilities\") pod \"redhat-operators-6nmxg\" (UID: \"279f32be-baa9-4cf7-a540-f6f8552c2096\") " pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:16 crc kubenswrapper[4814]: I0227 16:37:16.328479 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/279f32be-baa9-4cf7-a540-f6f8552c2096-catalog-content\") pod \"redhat-operators-6nmxg\" (UID: \"279f32be-baa9-4cf7-a540-f6f8552c2096\") " pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:16 crc kubenswrapper[4814]: I0227 16:37:16.329133 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/279f32be-baa9-4cf7-a540-f6f8552c2096-catalog-content\") pod \"redhat-operators-6nmxg\" (UID: \"279f32be-baa9-4cf7-a540-f6f8552c2096\") " pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:16 crc kubenswrapper[4814]: I0227 16:37:16.329170 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/279f32be-baa9-4cf7-a540-f6f8552c2096-utilities\") pod \"redhat-operators-6nmxg\" (UID: \"279f32be-baa9-4cf7-a540-f6f8552c2096\") " pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:16 crc kubenswrapper[4814]: I0227 16:37:16.354529 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2zc4\" (UniqueName: \"kubernetes.io/projected/279f32be-baa9-4cf7-a540-f6f8552c2096-kube-api-access-s2zc4\") pod \"redhat-operators-6nmxg\" (UID: \"279f32be-baa9-4cf7-a540-f6f8552c2096\") " pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:16 crc kubenswrapper[4814]: I0227 16:37:16.493970 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:16 crc kubenswrapper[4814]: I0227 16:37:16.759732 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6nmxg"] Feb 27 16:37:16 crc kubenswrapper[4814]: W0227 16:37:16.774417 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod279f32be_baa9_4cf7_a540_f6f8552c2096.slice/crio-c57818078a5b12218bb8ad3b425ca7bbe4a25f4791f7811e8930737480c13ee4 WatchSource:0}: Error finding container c57818078a5b12218bb8ad3b425ca7bbe4a25f4791f7811e8930737480c13ee4: Status 404 returned error can't find the container with id c57818078a5b12218bb8ad3b425ca7bbe4a25f4791f7811e8930737480c13ee4 Feb 27 16:37:16 crc kubenswrapper[4814]: I0227 16:37:16.855103 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6nmxg" event={"ID":"279f32be-baa9-4cf7-a540-f6f8552c2096","Type":"ContainerStarted","Data":"c57818078a5b12218bb8ad3b425ca7bbe4a25f4791f7811e8930737480c13ee4"} Feb 27 16:37:17 crc kubenswrapper[4814]: I0227 16:37:17.862585 4814 generic.go:334] "Generic (PLEG): container finished" podID="279f32be-baa9-4cf7-a540-f6f8552c2096" containerID="cb17c9698a9b5e8811b9b09aa304872540f2c25e919a4d378598fd17ead2ef3e" exitCode=0 Feb 27 16:37:17 crc kubenswrapper[4814]: I0227 16:37:17.864333 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6nmxg" event={"ID":"279f32be-baa9-4cf7-a540-f6f8552c2096","Type":"ContainerDied","Data":"cb17c9698a9b5e8811b9b09aa304872540f2c25e919a4d378598fd17ead2ef3e"} Feb 27 16:37:18 crc kubenswrapper[4814]: I0227 16:37:18.887679 4814 generic.go:334] "Generic (PLEG): container finished" podID="4fb39e04-954d-4a82-9d19-ece0f11c5118" containerID="be5059dc0d20b2183c12aa24d6e7e26aa69434d630e4dbc30d3a768eb8607042" exitCode=0 Feb 27 16:37:18 crc kubenswrapper[4814]: I0227 16:37:18.888051 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" event={"ID":"4fb39e04-954d-4a82-9d19-ece0f11c5118","Type":"ContainerDied","Data":"be5059dc0d20b2183c12aa24d6e7e26aa69434d630e4dbc30d3a768eb8607042"} Feb 27 16:37:19 crc kubenswrapper[4814]: I0227 16:37:19.913319 4814 generic.go:334] "Generic (PLEG): container finished" podID="4fb39e04-954d-4a82-9d19-ece0f11c5118" containerID="73d53ad63662c825230a89984c6b39f0a7ae1b311c078aa19d8caefafdbd2fcd" exitCode=0 Feb 27 16:37:19 crc kubenswrapper[4814]: I0227 16:37:19.913499 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" event={"ID":"4fb39e04-954d-4a82-9d19-ece0f11c5118","Type":"ContainerDied","Data":"73d53ad63662c825230a89984c6b39f0a7ae1b311c078aa19d8caefafdbd2fcd"} Feb 27 16:37:19 crc kubenswrapper[4814]: I0227 16:37:19.916637 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6nmxg" event={"ID":"279f32be-baa9-4cf7-a540-f6f8552c2096","Type":"ContainerStarted","Data":"939a5b115173f39ef6bcad9a2ce834bcc8e4e525ba89fd98afe40d9e16462527"} Feb 27 16:37:20 crc kubenswrapper[4814]: I0227 16:37:20.929942 4814 generic.go:334] "Generic (PLEG): container finished" podID="279f32be-baa9-4cf7-a540-f6f8552c2096" containerID="939a5b115173f39ef6bcad9a2ce834bcc8e4e525ba89fd98afe40d9e16462527" exitCode=0 Feb 27 16:37:20 crc kubenswrapper[4814]: I0227 16:37:20.930491 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6nmxg" event={"ID":"279f32be-baa9-4cf7-a540-f6f8552c2096","Type":"ContainerDied","Data":"939a5b115173f39ef6bcad9a2ce834bcc8e4e525ba89fd98afe40d9e16462527"} Feb 27 16:37:21 crc kubenswrapper[4814]: I0227 16:37:21.325384 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" Feb 27 16:37:21 crc kubenswrapper[4814]: I0227 16:37:21.409092 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4fb39e04-954d-4a82-9d19-ece0f11c5118-util\") pod \"4fb39e04-954d-4a82-9d19-ece0f11c5118\" (UID: \"4fb39e04-954d-4a82-9d19-ece0f11c5118\") " Feb 27 16:37:21 crc kubenswrapper[4814]: I0227 16:37:21.433905 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fb39e04-954d-4a82-9d19-ece0f11c5118-util" (OuterVolumeSpecName: "util") pod "4fb39e04-954d-4a82-9d19-ece0f11c5118" (UID: "4fb39e04-954d-4a82-9d19-ece0f11c5118"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:37:21 crc kubenswrapper[4814]: I0227 16:37:21.510519 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4fb39e04-954d-4a82-9d19-ece0f11c5118-bundle\") pod \"4fb39e04-954d-4a82-9d19-ece0f11c5118\" (UID: \"4fb39e04-954d-4a82-9d19-ece0f11c5118\") " Feb 27 16:37:21 crc kubenswrapper[4814]: I0227 16:37:21.510594 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2hm4\" (UniqueName: \"kubernetes.io/projected/4fb39e04-954d-4a82-9d19-ece0f11c5118-kube-api-access-x2hm4\") pod \"4fb39e04-954d-4a82-9d19-ece0f11c5118\" (UID: \"4fb39e04-954d-4a82-9d19-ece0f11c5118\") " Feb 27 16:37:21 crc kubenswrapper[4814]: I0227 16:37:21.510816 4814 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4fb39e04-954d-4a82-9d19-ece0f11c5118-util\") on node \"crc\" DevicePath \"\"" Feb 27 16:37:21 crc kubenswrapper[4814]: I0227 16:37:21.511156 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fb39e04-954d-4a82-9d19-ece0f11c5118-bundle" (OuterVolumeSpecName: "bundle") pod "4fb39e04-954d-4a82-9d19-ece0f11c5118" (UID: "4fb39e04-954d-4a82-9d19-ece0f11c5118"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:37:21 crc kubenswrapper[4814]: I0227 16:37:21.523525 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fb39e04-954d-4a82-9d19-ece0f11c5118-kube-api-access-x2hm4" (OuterVolumeSpecName: "kube-api-access-x2hm4") pod "4fb39e04-954d-4a82-9d19-ece0f11c5118" (UID: "4fb39e04-954d-4a82-9d19-ece0f11c5118"). InnerVolumeSpecName "kube-api-access-x2hm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:37:21 crc kubenswrapper[4814]: I0227 16:37:21.611926 4814 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4fb39e04-954d-4a82-9d19-ece0f11c5118-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:37:21 crc kubenswrapper[4814]: I0227 16:37:21.611959 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2hm4\" (UniqueName: \"kubernetes.io/projected/4fb39e04-954d-4a82-9d19-ece0f11c5118-kube-api-access-x2hm4\") on node \"crc\" DevicePath \"\"" Feb 27 16:37:21 crc kubenswrapper[4814]: I0227 16:37:21.941245 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" event={"ID":"4fb39e04-954d-4a82-9d19-ece0f11c5118","Type":"ContainerDied","Data":"a5376bc0efae9864a913d2b5136864a0c1e3e16e1b7a65f411edfa9eddda1784"} Feb 27 16:37:21 crc kubenswrapper[4814]: I0227 16:37:21.941359 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5376bc0efae9864a913d2b5136864a0c1e3e16e1b7a65f411edfa9eddda1784" Feb 27 16:37:21 crc kubenswrapper[4814]: I0227 16:37:21.941302 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr" Feb 27 16:37:21 crc kubenswrapper[4814]: I0227 16:37:21.945718 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6nmxg" event={"ID":"279f32be-baa9-4cf7-a540-f6f8552c2096","Type":"ContainerStarted","Data":"d55c97ed3c41f0e468f9ad663a39b7882f411f90e2839331726692d718cabab2"} Feb 27 16:37:21 crc kubenswrapper[4814]: I0227 16:37:21.971831 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6nmxg" podStartSLOduration=2.83866549 podStartE2EDuration="5.971785547s" podCreationTimestamp="2026-02-27 16:37:16 +0000 UTC" firstStartedPulling="2026-02-27 16:37:18.320635016 +0000 UTC m=+850.773259886" lastFinishedPulling="2026-02-27 16:37:21.453755113 +0000 UTC m=+853.906379943" observedRunningTime="2026-02-27 16:37:21.971636723 +0000 UTC m=+854.424261553" watchObservedRunningTime="2026-02-27 16:37:21.971785547 +0000 UTC m=+854.424410387" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.309679 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-75c5dccd6c-r6vt9"] Feb 27 16:37:25 crc kubenswrapper[4814]: E0227 16:37:25.310440 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fb39e04-954d-4a82-9d19-ece0f11c5118" containerName="pull" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.310459 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fb39e04-954d-4a82-9d19-ece0f11c5118" containerName="pull" Feb 27 16:37:25 crc kubenswrapper[4814]: E0227 16:37:25.310479 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fb39e04-954d-4a82-9d19-ece0f11c5118" containerName="extract" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.310486 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fb39e04-954d-4a82-9d19-ece0f11c5118" containerName="extract" Feb 27 16:37:25 crc kubenswrapper[4814]: E0227 16:37:25.310508 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fb39e04-954d-4a82-9d19-ece0f11c5118" containerName="util" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.310516 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fb39e04-954d-4a82-9d19-ece0f11c5118" containerName="util" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.310645 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fb39e04-954d-4a82-9d19-ece0f11c5118" containerName="extract" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.311214 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-r6vt9" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.313540 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.314040 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.319374 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-4sw54" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.321195 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-75c5dccd6c-r6vt9"] Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.377727 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5lxp\" (UniqueName: \"kubernetes.io/projected/154b3bf1-f2f5-4e4b-9110-b2097784f5d8-kube-api-access-k5lxp\") pod \"nmstate-operator-75c5dccd6c-r6vt9\" (UID: \"154b3bf1-f2f5-4e4b-9110-b2097784f5d8\") " pod="openshift-nmstate/nmstate-operator-75c5dccd6c-r6vt9" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.479241 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5lxp\" (UniqueName: \"kubernetes.io/projected/154b3bf1-f2f5-4e4b-9110-b2097784f5d8-kube-api-access-k5lxp\") pod \"nmstate-operator-75c5dccd6c-r6vt9\" (UID: \"154b3bf1-f2f5-4e4b-9110-b2097784f5d8\") " pod="openshift-nmstate/nmstate-operator-75c5dccd6c-r6vt9" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.515531 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5lxp\" (UniqueName: \"kubernetes.io/projected/154b3bf1-f2f5-4e4b-9110-b2097784f5d8-kube-api-access-k5lxp\") pod \"nmstate-operator-75c5dccd6c-r6vt9\" (UID: \"154b3bf1-f2f5-4e4b-9110-b2097784f5d8\") " pod="openshift-nmstate/nmstate-operator-75c5dccd6c-r6vt9" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.630717 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-r6vt9" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.784806 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n275h"] Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.792630 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.795980 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n275h"] Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.884976 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bfj8\" (UniqueName: \"kubernetes.io/projected/9e9625fe-2036-4308-a07f-0548705c91dc-kube-api-access-6bfj8\") pod \"redhat-marketplace-n275h\" (UID: \"9e9625fe-2036-4308-a07f-0548705c91dc\") " pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.885074 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e9625fe-2036-4308-a07f-0548705c91dc-catalog-content\") pod \"redhat-marketplace-n275h\" (UID: \"9e9625fe-2036-4308-a07f-0548705c91dc\") " pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.885136 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e9625fe-2036-4308-a07f-0548705c91dc-utilities\") pod \"redhat-marketplace-n275h\" (UID: \"9e9625fe-2036-4308-a07f-0548705c91dc\") " pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.987375 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e9625fe-2036-4308-a07f-0548705c91dc-utilities\") pod \"redhat-marketplace-n275h\" (UID: \"9e9625fe-2036-4308-a07f-0548705c91dc\") " pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.987469 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bfj8\" (UniqueName: \"kubernetes.io/projected/9e9625fe-2036-4308-a07f-0548705c91dc-kube-api-access-6bfj8\") pod \"redhat-marketplace-n275h\" (UID: \"9e9625fe-2036-4308-a07f-0548705c91dc\") " pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.987513 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e9625fe-2036-4308-a07f-0548705c91dc-catalog-content\") pod \"redhat-marketplace-n275h\" (UID: \"9e9625fe-2036-4308-a07f-0548705c91dc\") " pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.988284 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e9625fe-2036-4308-a07f-0548705c91dc-catalog-content\") pod \"redhat-marketplace-n275h\" (UID: \"9e9625fe-2036-4308-a07f-0548705c91dc\") " pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:25 crc kubenswrapper[4814]: I0227 16:37:25.988368 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e9625fe-2036-4308-a07f-0548705c91dc-utilities\") pod \"redhat-marketplace-n275h\" (UID: \"9e9625fe-2036-4308-a07f-0548705c91dc\") " pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:26 crc kubenswrapper[4814]: I0227 16:37:26.006493 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bfj8\" (UniqueName: \"kubernetes.io/projected/9e9625fe-2036-4308-a07f-0548705c91dc-kube-api-access-6bfj8\") pod \"redhat-marketplace-n275h\" (UID: \"9e9625fe-2036-4308-a07f-0548705c91dc\") " pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:26 crc kubenswrapper[4814]: I0227 16:37:26.144480 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:26 crc kubenswrapper[4814]: I0227 16:37:26.216412 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-75c5dccd6c-r6vt9"] Feb 27 16:37:26 crc kubenswrapper[4814]: W0227 16:37:26.246567 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod154b3bf1_f2f5_4e4b_9110_b2097784f5d8.slice/crio-5a836607988a4b49d93c00bf6b51205308387ec6e68fbd0e958332a6ad4b8ebd WatchSource:0}: Error finding container 5a836607988a4b49d93c00bf6b51205308387ec6e68fbd0e958332a6ad4b8ebd: Status 404 returned error can't find the container with id 5a836607988a4b49d93c00bf6b51205308387ec6e68fbd0e958332a6ad4b8ebd Feb 27 16:37:26 crc kubenswrapper[4814]: I0227 16:37:26.416524 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n275h"] Feb 27 16:37:26 crc kubenswrapper[4814]: W0227 16:37:26.422122 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e9625fe_2036_4308_a07f_0548705c91dc.slice/crio-e17f7ac292a2c9554e25f86b9e190dbfc6fd3bbd9e08ac18a4fc6ec758b52de5 WatchSource:0}: Error finding container e17f7ac292a2c9554e25f86b9e190dbfc6fd3bbd9e08ac18a4fc6ec758b52de5: Status 404 returned error can't find the container with id e17f7ac292a2c9554e25f86b9e190dbfc6fd3bbd9e08ac18a4fc6ec758b52de5 Feb 27 16:37:26 crc kubenswrapper[4814]: I0227 16:37:26.496292 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:26 crc kubenswrapper[4814]: I0227 16:37:26.496331 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:26 crc kubenswrapper[4814]: I0227 16:37:26.980620 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n275h" event={"ID":"9e9625fe-2036-4308-a07f-0548705c91dc","Type":"ContainerStarted","Data":"e17f7ac292a2c9554e25f86b9e190dbfc6fd3bbd9e08ac18a4fc6ec758b52de5"} Feb 27 16:37:26 crc kubenswrapper[4814]: I0227 16:37:26.982380 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-r6vt9" event={"ID":"154b3bf1-f2f5-4e4b-9110-b2097784f5d8","Type":"ContainerStarted","Data":"5a836607988a4b49d93c00bf6b51205308387ec6e68fbd0e958332a6ad4b8ebd"} Feb 27 16:37:27 crc kubenswrapper[4814]: I0227 16:37:27.543505 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6nmxg" podUID="279f32be-baa9-4cf7-a540-f6f8552c2096" containerName="registry-server" probeResult="failure" output=< Feb 27 16:37:27 crc kubenswrapper[4814]: timeout: failed to connect service ":50051" within 1s Feb 27 16:37:27 crc kubenswrapper[4814]: > Feb 27 16:37:27 crc kubenswrapper[4814]: I0227 16:37:27.989770 4814 generic.go:334] "Generic (PLEG): container finished" podID="9e9625fe-2036-4308-a07f-0548705c91dc" containerID="6784a11cc69dcdc5babbb8587209880debe22e3cf9c61c2d69ebe26b82bae460" exitCode=0 Feb 27 16:37:27 crc kubenswrapper[4814]: I0227 16:37:27.989818 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n275h" event={"ID":"9e9625fe-2036-4308-a07f-0548705c91dc","Type":"ContainerDied","Data":"6784a11cc69dcdc5babbb8587209880debe22e3cf9c61c2d69ebe26b82bae460"} Feb 27 16:37:28 crc kubenswrapper[4814]: I0227 16:37:28.999069 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-r6vt9" event={"ID":"154b3bf1-f2f5-4e4b-9110-b2097784f5d8","Type":"ContainerStarted","Data":"e10fc94e7b28fa9c4f9d65375c0ce5dbf15944b5c34be2efaa6dae3a7d11de05"} Feb 27 16:37:29 crc kubenswrapper[4814]: I0227 16:37:29.031453 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-r6vt9" podStartSLOduration=1.673277998 podStartE2EDuration="4.031422789s" podCreationTimestamp="2026-02-27 16:37:25 +0000 UTC" firstStartedPulling="2026-02-27 16:37:26.249979403 +0000 UTC m=+858.702604243" lastFinishedPulling="2026-02-27 16:37:28.608124204 +0000 UTC m=+861.060749034" observedRunningTime="2026-02-27 16:37:29.027052238 +0000 UTC m=+861.479677108" watchObservedRunningTime="2026-02-27 16:37:29.031422789 +0000 UTC m=+861.484047649" Feb 27 16:37:29 crc kubenswrapper[4814]: E0227 16:37:29.545430 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e9625fe_2036_4308_a07f_0548705c91dc.slice/crio-927158e6a433a614bf7d0ecbe61d1dc774f5efe22f20e4e6356ce727d11a425d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e9625fe_2036_4308_a07f_0548705c91dc.slice/crio-conmon-927158e6a433a614bf7d0ecbe61d1dc774f5efe22f20e4e6356ce727d11a425d.scope\": RecentStats: unable to find data in memory cache]" Feb 27 16:37:30 crc kubenswrapper[4814]: I0227 16:37:30.010871 4814 generic.go:334] "Generic (PLEG): container finished" podID="9e9625fe-2036-4308-a07f-0548705c91dc" containerID="927158e6a433a614bf7d0ecbe61d1dc774f5efe22f20e4e6356ce727d11a425d" exitCode=0 Feb 27 16:37:30 crc kubenswrapper[4814]: I0227 16:37:30.011016 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n275h" event={"ID":"9e9625fe-2036-4308-a07f-0548705c91dc","Type":"ContainerDied","Data":"927158e6a433a614bf7d0ecbe61d1dc774f5efe22f20e4e6356ce727d11a425d"} Feb 27 16:37:31 crc kubenswrapper[4814]: I0227 16:37:31.029214 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n275h" event={"ID":"9e9625fe-2036-4308-a07f-0548705c91dc","Type":"ContainerStarted","Data":"fb7d6b6f669010deeb79746a3a9322ba7fe2ccf0a193dc249340379d744a69d2"} Feb 27 16:37:31 crc kubenswrapper[4814]: I0227 16:37:31.054638 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n275h" podStartSLOduration=4.149899405 podStartE2EDuration="6.054604547s" podCreationTimestamp="2026-02-27 16:37:25 +0000 UTC" firstStartedPulling="2026-02-27 16:37:28.549188002 +0000 UTC m=+861.001812842" lastFinishedPulling="2026-02-27 16:37:30.453893154 +0000 UTC m=+862.906517984" observedRunningTime="2026-02-27 16:37:31.051045969 +0000 UTC m=+863.503670839" watchObservedRunningTime="2026-02-27 16:37:31.054604547 +0000 UTC m=+863.507229407" Feb 27 16:37:34 crc kubenswrapper[4814]: I0227 16:37:34.852468 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-69594cc75-9xknj"] Feb 27 16:37:34 crc kubenswrapper[4814]: I0227 16:37:34.853734 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-69594cc75-9xknj" Feb 27 16:37:34 crc kubenswrapper[4814]: I0227 16:37:34.856163 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-6nwx2" Feb 27 16:37:34 crc kubenswrapper[4814]: I0227 16:37:34.870459 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-69594cc75-9xknj"] Feb 27 16:37:34 crc kubenswrapper[4814]: I0227 16:37:34.875270 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-786f45cff4-bbbjp"] Feb 27 16:37:34 crc kubenswrapper[4814]: I0227 16:37:34.876324 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-786f45cff4-bbbjp" Feb 27 16:37:34 crc kubenswrapper[4814]: I0227 16:37:34.900918 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 27 16:37:34 crc kubenswrapper[4814]: I0227 16:37:34.910139 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-786f45cff4-bbbjp"] Feb 27 16:37:34 crc kubenswrapper[4814]: I0227 16:37:34.926532 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-6j6pw"] Feb 27 16:37:34 crc kubenswrapper[4814]: I0227 16:37:34.927543 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6j6pw" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.010768 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn"] Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.011859 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.014668 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.014902 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.017640 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-jpwlt" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.019856 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwwv6\" (UniqueName: \"kubernetes.io/projected/6df39380-a69c-4ef3-94ba-6a5ef31e7f50-kube-api-access-hwwv6\") pod \"nmstate-metrics-69594cc75-9xknj\" (UID: \"6df39380-a69c-4ef3-94ba-6a5ef31e7f50\") " pod="openshift-nmstate/nmstate-metrics-69594cc75-9xknj" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.019981 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/65270046-4e69-4b5c-b07f-3f401949f32b-tls-key-pair\") pod \"nmstate-webhook-786f45cff4-bbbjp\" (UID: \"65270046-4e69-4b5c-b07f-3f401949f32b\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-bbbjp" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.020085 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58hf9\" (UniqueName: \"kubernetes.io/projected/65270046-4e69-4b5c-b07f-3f401949f32b-kube-api-access-58hf9\") pod \"nmstate-webhook-786f45cff4-bbbjp\" (UID: \"65270046-4e69-4b5c-b07f-3f401949f32b\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-bbbjp" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.024136 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn"] Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.121240 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/694e9d8b-01d0-444c-bf6f-cc4c3cd4f542-dbus-socket\") pod \"nmstate-handler-6j6pw\" (UID: \"694e9d8b-01d0-444c-bf6f-cc4c3cd4f542\") " pod="openshift-nmstate/nmstate-handler-6j6pw" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.121340 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlprj\" (UniqueName: \"kubernetes.io/projected/ec633892-636c-460c-8c7a-1b594ba270b5-kube-api-access-qlprj\") pod \"nmstate-console-plugin-5dcbbd79cf-5hqsn\" (UID: \"ec633892-636c-460c-8c7a-1b594ba270b5\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.121388 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwwv6\" (UniqueName: \"kubernetes.io/projected/6df39380-a69c-4ef3-94ba-6a5ef31e7f50-kube-api-access-hwwv6\") pod \"nmstate-metrics-69594cc75-9xknj\" (UID: \"6df39380-a69c-4ef3-94ba-6a5ef31e7f50\") " pod="openshift-nmstate/nmstate-metrics-69594cc75-9xknj" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.121430 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/65270046-4e69-4b5c-b07f-3f401949f32b-tls-key-pair\") pod \"nmstate-webhook-786f45cff4-bbbjp\" (UID: \"65270046-4e69-4b5c-b07f-3f401949f32b\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-bbbjp" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.121509 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58hf9\" (UniqueName: \"kubernetes.io/projected/65270046-4e69-4b5c-b07f-3f401949f32b-kube-api-access-58hf9\") pod \"nmstate-webhook-786f45cff4-bbbjp\" (UID: \"65270046-4e69-4b5c-b07f-3f401949f32b\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-bbbjp" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.121556 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll85r\" (UniqueName: \"kubernetes.io/projected/694e9d8b-01d0-444c-bf6f-cc4c3cd4f542-kube-api-access-ll85r\") pod \"nmstate-handler-6j6pw\" (UID: \"694e9d8b-01d0-444c-bf6f-cc4c3cd4f542\") " pod="openshift-nmstate/nmstate-handler-6j6pw" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.121587 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ec633892-636c-460c-8c7a-1b594ba270b5-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-5hqsn\" (UID: \"ec633892-636c-460c-8c7a-1b594ba270b5\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.121638 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/694e9d8b-01d0-444c-bf6f-cc4c3cd4f542-nmstate-lock\") pod \"nmstate-handler-6j6pw\" (UID: \"694e9d8b-01d0-444c-bf6f-cc4c3cd4f542\") " pod="openshift-nmstate/nmstate-handler-6j6pw" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.121683 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ec633892-636c-460c-8c7a-1b594ba270b5-nginx-conf\") pod \"nmstate-console-plugin-5dcbbd79cf-5hqsn\" (UID: \"ec633892-636c-460c-8c7a-1b594ba270b5\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.121722 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/694e9d8b-01d0-444c-bf6f-cc4c3cd4f542-ovs-socket\") pod \"nmstate-handler-6j6pw\" (UID: \"694e9d8b-01d0-444c-bf6f-cc4c3cd4f542\") " pod="openshift-nmstate/nmstate-handler-6j6pw" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.133068 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/65270046-4e69-4b5c-b07f-3f401949f32b-tls-key-pair\") pod \"nmstate-webhook-786f45cff4-bbbjp\" (UID: \"65270046-4e69-4b5c-b07f-3f401949f32b\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-bbbjp" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.143051 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58hf9\" (UniqueName: \"kubernetes.io/projected/65270046-4e69-4b5c-b07f-3f401949f32b-kube-api-access-58hf9\") pod \"nmstate-webhook-786f45cff4-bbbjp\" (UID: \"65270046-4e69-4b5c-b07f-3f401949f32b\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-bbbjp" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.151793 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwwv6\" (UniqueName: \"kubernetes.io/projected/6df39380-a69c-4ef3-94ba-6a5ef31e7f50-kube-api-access-hwwv6\") pod \"nmstate-metrics-69594cc75-9xknj\" (UID: \"6df39380-a69c-4ef3-94ba-6a5ef31e7f50\") " pod="openshift-nmstate/nmstate-metrics-69594cc75-9xknj" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.176633 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-69594cc75-9xknj" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.204706 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-786f45cff4-bbbjp" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.222381 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/694e9d8b-01d0-444c-bf6f-cc4c3cd4f542-dbus-socket\") pod \"nmstate-handler-6j6pw\" (UID: \"694e9d8b-01d0-444c-bf6f-cc4c3cd4f542\") " pod="openshift-nmstate/nmstate-handler-6j6pw" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.222422 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlprj\" (UniqueName: \"kubernetes.io/projected/ec633892-636c-460c-8c7a-1b594ba270b5-kube-api-access-qlprj\") pod \"nmstate-console-plugin-5dcbbd79cf-5hqsn\" (UID: \"ec633892-636c-460c-8c7a-1b594ba270b5\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.222486 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll85r\" (UniqueName: \"kubernetes.io/projected/694e9d8b-01d0-444c-bf6f-cc4c3cd4f542-kube-api-access-ll85r\") pod \"nmstate-handler-6j6pw\" (UID: \"694e9d8b-01d0-444c-bf6f-cc4c3cd4f542\") " pod="openshift-nmstate/nmstate-handler-6j6pw" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.222505 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ec633892-636c-460c-8c7a-1b594ba270b5-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-5hqsn\" (UID: \"ec633892-636c-460c-8c7a-1b594ba270b5\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.222522 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/694e9d8b-01d0-444c-bf6f-cc4c3cd4f542-nmstate-lock\") pod \"nmstate-handler-6j6pw\" (UID: \"694e9d8b-01d0-444c-bf6f-cc4c3cd4f542\") " pod="openshift-nmstate/nmstate-handler-6j6pw" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.222542 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ec633892-636c-460c-8c7a-1b594ba270b5-nginx-conf\") pod \"nmstate-console-plugin-5dcbbd79cf-5hqsn\" (UID: \"ec633892-636c-460c-8c7a-1b594ba270b5\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.222560 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/694e9d8b-01d0-444c-bf6f-cc4c3cd4f542-ovs-socket\") pod \"nmstate-handler-6j6pw\" (UID: \"694e9d8b-01d0-444c-bf6f-cc4c3cd4f542\") " pod="openshift-nmstate/nmstate-handler-6j6pw" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.222635 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/694e9d8b-01d0-444c-bf6f-cc4c3cd4f542-ovs-socket\") pod \"nmstate-handler-6j6pw\" (UID: \"694e9d8b-01d0-444c-bf6f-cc4c3cd4f542\") " pod="openshift-nmstate/nmstate-handler-6j6pw" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.224324 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ec633892-636c-460c-8c7a-1b594ba270b5-nginx-conf\") pod \"nmstate-console-plugin-5dcbbd79cf-5hqsn\" (UID: \"ec633892-636c-460c-8c7a-1b594ba270b5\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.224560 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/694e9d8b-01d0-444c-bf6f-cc4c3cd4f542-dbus-socket\") pod \"nmstate-handler-6j6pw\" (UID: \"694e9d8b-01d0-444c-bf6f-cc4c3cd4f542\") " pod="openshift-nmstate/nmstate-handler-6j6pw" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.226356 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/694e9d8b-01d0-444c-bf6f-cc4c3cd4f542-nmstate-lock\") pod \"nmstate-handler-6j6pw\" (UID: \"694e9d8b-01d0-444c-bf6f-cc4c3cd4f542\") " pod="openshift-nmstate/nmstate-handler-6j6pw" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.231536 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ec633892-636c-460c-8c7a-1b594ba270b5-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-5hqsn\" (UID: \"ec633892-636c-460c-8c7a-1b594ba270b5\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.242289 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7fcd9679d-dtpk9"] Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.242995 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.248412 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlprj\" (UniqueName: \"kubernetes.io/projected/ec633892-636c-460c-8c7a-1b594ba270b5-kube-api-access-qlprj\") pod \"nmstate-console-plugin-5dcbbd79cf-5hqsn\" (UID: \"ec633892-636c-460c-8c7a-1b594ba270b5\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.250133 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll85r\" (UniqueName: \"kubernetes.io/projected/694e9d8b-01d0-444c-bf6f-cc4c3cd4f542-kube-api-access-ll85r\") pod \"nmstate-handler-6j6pw\" (UID: \"694e9d8b-01d0-444c-bf6f-cc4c3cd4f542\") " pod="openshift-nmstate/nmstate-handler-6j6pw" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.260918 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7fcd9679d-dtpk9"] Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.331639 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.425302 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdsmj\" (UniqueName: \"kubernetes.io/projected/66be49e1-0fb0-4a07-9683-2a5df8045f90-kube-api-access-kdsmj\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.425364 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/66be49e1-0fb0-4a07-9683-2a5df8045f90-console-serving-cert\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.425442 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66be49e1-0fb0-4a07-9683-2a5df8045f90-trusted-ca-bundle\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.425487 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/66be49e1-0fb0-4a07-9683-2a5df8045f90-console-config\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.425718 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66be49e1-0fb0-4a07-9683-2a5df8045f90-service-ca\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.425746 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/66be49e1-0fb0-4a07-9683-2a5df8045f90-console-oauth-config\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.425876 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/66be49e1-0fb0-4a07-9683-2a5df8045f90-oauth-serving-cert\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.440624 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-69594cc75-9xknj"] Feb 27 16:37:35 crc kubenswrapper[4814]: W0227 16:37:35.462181 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6df39380_a69c_4ef3_94ba_6a5ef31e7f50.slice/crio-c6c8523d55218790efc63a3d84d851697cd6f77c5895bc2486355a816fc05000 WatchSource:0}: Error finding container c6c8523d55218790efc63a3d84d851697cd6f77c5895bc2486355a816fc05000: Status 404 returned error can't find the container with id c6c8523d55218790efc63a3d84d851697cd6f77c5895bc2486355a816fc05000 Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.518921 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-786f45cff4-bbbjp"] Feb 27 16:37:35 crc kubenswrapper[4814]: W0227 16:37:35.525931 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65270046_4e69_4b5c_b07f_3f401949f32b.slice/crio-b3b2bc3b52c7fd2e6490a373e4850a431ae21c286aee45e01d83d0b23f325d6f WatchSource:0}: Error finding container b3b2bc3b52c7fd2e6490a373e4850a431ae21c286aee45e01d83d0b23f325d6f: Status 404 returned error can't find the container with id b3b2bc3b52c7fd2e6490a373e4850a431ae21c286aee45e01d83d0b23f325d6f Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.527024 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/66be49e1-0fb0-4a07-9683-2a5df8045f90-oauth-serving-cert\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.527114 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdsmj\" (UniqueName: \"kubernetes.io/projected/66be49e1-0fb0-4a07-9683-2a5df8045f90-kube-api-access-kdsmj\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.527168 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/66be49e1-0fb0-4a07-9683-2a5df8045f90-console-serving-cert\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.527204 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66be49e1-0fb0-4a07-9683-2a5df8045f90-trusted-ca-bundle\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.527281 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/66be49e1-0fb0-4a07-9683-2a5df8045f90-console-config\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.527303 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66be49e1-0fb0-4a07-9683-2a5df8045f90-service-ca\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.527329 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/66be49e1-0fb0-4a07-9683-2a5df8045f90-console-oauth-config\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.527939 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/66be49e1-0fb0-4a07-9683-2a5df8045f90-oauth-serving-cert\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.528297 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/66be49e1-0fb0-4a07-9683-2a5df8045f90-console-config\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.528359 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66be49e1-0fb0-4a07-9683-2a5df8045f90-trusted-ca-bundle\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.528545 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66be49e1-0fb0-4a07-9683-2a5df8045f90-service-ca\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.534908 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/66be49e1-0fb0-4a07-9683-2a5df8045f90-console-serving-cert\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.536497 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/66be49e1-0fb0-4a07-9683-2a5df8045f90-console-oauth-config\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.545953 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdsmj\" (UniqueName: \"kubernetes.io/projected/66be49e1-0fb0-4a07-9683-2a5df8045f90-kube-api-access-kdsmj\") pod \"console-7fcd9679d-dtpk9\" (UID: \"66be49e1-0fb0-4a07-9683-2a5df8045f90\") " pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.548581 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6j6pw" Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.558356 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn"] Feb 27 16:37:35 crc kubenswrapper[4814]: W0227 16:37:35.570620 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod694e9d8b_01d0_444c_bf6f_cc4c3cd4f542.slice/crio-532e851faf182efc4f391d5d2711916a602e323dcbfd27c19e9211fbf61e2101 WatchSource:0}: Error finding container 532e851faf182efc4f391d5d2711916a602e323dcbfd27c19e9211fbf61e2101: Status 404 returned error can't find the container with id 532e851faf182efc4f391d5d2711916a602e323dcbfd27c19e9211fbf61e2101 Feb 27 16:37:35 crc kubenswrapper[4814]: I0227 16:37:35.595449 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:36 crc kubenswrapper[4814]: I0227 16:37:36.069222 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6j6pw" event={"ID":"694e9d8b-01d0-444c-bf6f-cc4c3cd4f542","Type":"ContainerStarted","Data":"532e851faf182efc4f391d5d2711916a602e323dcbfd27c19e9211fbf61e2101"} Feb 27 16:37:36 crc kubenswrapper[4814]: I0227 16:37:36.071303 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn" event={"ID":"ec633892-636c-460c-8c7a-1b594ba270b5","Type":"ContainerStarted","Data":"8c01b90a434d3b8695366d3c851809e8f1ae482ad4e2d35c0088a886f69c7ef8"} Feb 27 16:37:36 crc kubenswrapper[4814]: I0227 16:37:36.072870 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-69594cc75-9xknj" event={"ID":"6df39380-a69c-4ef3-94ba-6a5ef31e7f50","Type":"ContainerStarted","Data":"c6c8523d55218790efc63a3d84d851697cd6f77c5895bc2486355a816fc05000"} Feb 27 16:37:36 crc kubenswrapper[4814]: I0227 16:37:36.073999 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-786f45cff4-bbbjp" event={"ID":"65270046-4e69-4b5c-b07f-3f401949f32b","Type":"ContainerStarted","Data":"b3b2bc3b52c7fd2e6490a373e4850a431ae21c286aee45e01d83d0b23f325d6f"} Feb 27 16:37:36 crc kubenswrapper[4814]: I0227 16:37:36.074932 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7fcd9679d-dtpk9"] Feb 27 16:37:36 crc kubenswrapper[4814]: W0227 16:37:36.081800 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66be49e1_0fb0_4a07_9683_2a5df8045f90.slice/crio-26174e38898d3595df7a3227dcac70ea250043cdd71c85fb967ed068951c7631 WatchSource:0}: Error finding container 26174e38898d3595df7a3227dcac70ea250043cdd71c85fb967ed068951c7631: Status 404 returned error can't find the container with id 26174e38898d3595df7a3227dcac70ea250043cdd71c85fb967ed068951c7631 Feb 27 16:37:36 crc kubenswrapper[4814]: I0227 16:37:36.145709 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:36 crc kubenswrapper[4814]: I0227 16:37:36.146616 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:36 crc kubenswrapper[4814]: I0227 16:37:36.232159 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:36 crc kubenswrapper[4814]: I0227 16:37:36.544448 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:36 crc kubenswrapper[4814]: I0227 16:37:36.606340 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:37 crc kubenswrapper[4814]: I0227 16:37:37.083920 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7fcd9679d-dtpk9" event={"ID":"66be49e1-0fb0-4a07-9683-2a5df8045f90","Type":"ContainerStarted","Data":"ed3aa5e28966e6eedd15a0d5f3cba509c30cb5e819cf0db1fa30d33097d18b6b"} Feb 27 16:37:37 crc kubenswrapper[4814]: I0227 16:37:37.083998 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7fcd9679d-dtpk9" event={"ID":"66be49e1-0fb0-4a07-9683-2a5df8045f90","Type":"ContainerStarted","Data":"26174e38898d3595df7a3227dcac70ea250043cdd71c85fb967ed068951c7631"} Feb 27 16:37:37 crc kubenswrapper[4814]: I0227 16:37:37.103795 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7fcd9679d-dtpk9" podStartSLOduration=2.103770651 podStartE2EDuration="2.103770651s" podCreationTimestamp="2026-02-27 16:37:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:37:37.103505223 +0000 UTC m=+869.556130043" watchObservedRunningTime="2026-02-27 16:37:37.103770651 +0000 UTC m=+869.556395481" Feb 27 16:37:37 crc kubenswrapper[4814]: I0227 16:37:37.167019 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:39 crc kubenswrapper[4814]: I0227 16:37:39.102132 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-69594cc75-9xknj" event={"ID":"6df39380-a69c-4ef3-94ba-6a5ef31e7f50","Type":"ContainerStarted","Data":"762dd6735525ea5e583c85e2e88aef2981aa109d46a9cc9f64fcce3b7bc2982d"} Feb 27 16:37:39 crc kubenswrapper[4814]: I0227 16:37:39.104400 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-786f45cff4-bbbjp" event={"ID":"65270046-4e69-4b5c-b07f-3f401949f32b","Type":"ContainerStarted","Data":"1efccfd9946caa049eb76ba43241c7b50651147c1a778676332285476ba1bedc"} Feb 27 16:37:39 crc kubenswrapper[4814]: I0227 16:37:39.105605 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-786f45cff4-bbbjp" Feb 27 16:37:39 crc kubenswrapper[4814]: I0227 16:37:39.108146 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6j6pw" event={"ID":"694e9d8b-01d0-444c-bf6f-cc4c3cd4f542","Type":"ContainerStarted","Data":"98ed72db30e4c319b75f142496276e4087127971ff2b4afb0133c7992c8287b7"} Feb 27 16:37:39 crc kubenswrapper[4814]: I0227 16:37:39.108370 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-6j6pw" Feb 27 16:37:39 crc kubenswrapper[4814]: I0227 16:37:39.112467 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn" event={"ID":"ec633892-636c-460c-8c7a-1b594ba270b5","Type":"ContainerStarted","Data":"b730b45f543a5e6cd5aa67837a6d18d9e9a3b113c0b1e7333760bfa38257f124"} Feb 27 16:37:39 crc kubenswrapper[4814]: I0227 16:37:39.134923 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-786f45cff4-bbbjp" podStartSLOduration=2.168132923 podStartE2EDuration="5.134884294s" podCreationTimestamp="2026-02-27 16:37:34 +0000 UTC" firstStartedPulling="2026-02-27 16:37:35.532838801 +0000 UTC m=+867.985463631" lastFinishedPulling="2026-02-27 16:37:38.499590162 +0000 UTC m=+870.952215002" observedRunningTime="2026-02-27 16:37:39.127141352 +0000 UTC m=+871.579766222" watchObservedRunningTime="2026-02-27 16:37:39.134884294 +0000 UTC m=+871.587509164" Feb 27 16:37:39 crc kubenswrapper[4814]: I0227 16:37:39.152913 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-6j6pw" podStartSLOduration=2.254062918 podStartE2EDuration="5.152887646s" podCreationTimestamp="2026-02-27 16:37:34 +0000 UTC" firstStartedPulling="2026-02-27 16:37:35.573122423 +0000 UTC m=+868.025747253" lastFinishedPulling="2026-02-27 16:37:38.471947151 +0000 UTC m=+870.924571981" observedRunningTime="2026-02-27 16:37:39.152386631 +0000 UTC m=+871.605011491" watchObservedRunningTime="2026-02-27 16:37:39.152887646 +0000 UTC m=+871.605512486" Feb 27 16:37:39 crc kubenswrapper[4814]: I0227 16:37:39.176757 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-5hqsn" podStartSLOduration=2.274944466 podStartE2EDuration="5.176711743s" podCreationTimestamp="2026-02-27 16:37:34 +0000 UTC" firstStartedPulling="2026-02-27 16:37:35.56439966 +0000 UTC m=+868.017024490" lastFinishedPulling="2026-02-27 16:37:38.466166897 +0000 UTC m=+870.918791767" observedRunningTime="2026-02-27 16:37:39.169599269 +0000 UTC m=+871.622224109" watchObservedRunningTime="2026-02-27 16:37:39.176711743 +0000 UTC m=+871.629336583" Feb 27 16:37:39 crc kubenswrapper[4814]: I0227 16:37:39.354063 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n275h"] Feb 27 16:37:40 crc kubenswrapper[4814]: I0227 16:37:40.134556 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-n275h" podUID="9e9625fe-2036-4308-a07f-0548705c91dc" containerName="registry-server" containerID="cri-o://fb7d6b6f669010deeb79746a3a9322ba7fe2ccf0a193dc249340379d744a69d2" gracePeriod=2 Feb 27 16:37:40 crc kubenswrapper[4814]: I0227 16:37:40.159810 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6nmxg"] Feb 27 16:37:40 crc kubenswrapper[4814]: I0227 16:37:40.160169 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6nmxg" podUID="279f32be-baa9-4cf7-a540-f6f8552c2096" containerName="registry-server" containerID="cri-o://d55c97ed3c41f0e468f9ad663a39b7882f411f90e2839331726692d718cabab2" gracePeriod=2 Feb 27 16:37:40 crc kubenswrapper[4814]: I0227 16:37:40.615772 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:40 crc kubenswrapper[4814]: I0227 16:37:40.719326 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bfj8\" (UniqueName: \"kubernetes.io/projected/9e9625fe-2036-4308-a07f-0548705c91dc-kube-api-access-6bfj8\") pod \"9e9625fe-2036-4308-a07f-0548705c91dc\" (UID: \"9e9625fe-2036-4308-a07f-0548705c91dc\") " Feb 27 16:37:40 crc kubenswrapper[4814]: I0227 16:37:40.719410 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e9625fe-2036-4308-a07f-0548705c91dc-catalog-content\") pod \"9e9625fe-2036-4308-a07f-0548705c91dc\" (UID: \"9e9625fe-2036-4308-a07f-0548705c91dc\") " Feb 27 16:37:40 crc kubenswrapper[4814]: I0227 16:37:40.719655 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e9625fe-2036-4308-a07f-0548705c91dc-utilities\") pod \"9e9625fe-2036-4308-a07f-0548705c91dc\" (UID: \"9e9625fe-2036-4308-a07f-0548705c91dc\") " Feb 27 16:37:40 crc kubenswrapper[4814]: I0227 16:37:40.721005 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e9625fe-2036-4308-a07f-0548705c91dc-utilities" (OuterVolumeSpecName: "utilities") pod "9e9625fe-2036-4308-a07f-0548705c91dc" (UID: "9e9625fe-2036-4308-a07f-0548705c91dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:37:40 crc kubenswrapper[4814]: I0227 16:37:40.728999 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e9625fe-2036-4308-a07f-0548705c91dc-kube-api-access-6bfj8" (OuterVolumeSpecName: "kube-api-access-6bfj8") pod "9e9625fe-2036-4308-a07f-0548705c91dc" (UID: "9e9625fe-2036-4308-a07f-0548705c91dc"). InnerVolumeSpecName "kube-api-access-6bfj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:37:40 crc kubenswrapper[4814]: I0227 16:37:40.744212 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e9625fe-2036-4308-a07f-0548705c91dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e9625fe-2036-4308-a07f-0548705c91dc" (UID: "9e9625fe-2036-4308-a07f-0548705c91dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:37:40 crc kubenswrapper[4814]: I0227 16:37:40.822218 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e9625fe-2036-4308-a07f-0548705c91dc-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:37:40 crc kubenswrapper[4814]: I0227 16:37:40.822288 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e9625fe-2036-4308-a07f-0548705c91dc-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:37:40 crc kubenswrapper[4814]: I0227 16:37:40.822305 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bfj8\" (UniqueName: \"kubernetes.io/projected/9e9625fe-2036-4308-a07f-0548705c91dc-kube-api-access-6bfj8\") on node \"crc\" DevicePath \"\"" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.040622 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.143069 4814 generic.go:334] "Generic (PLEG): container finished" podID="279f32be-baa9-4cf7-a540-f6f8552c2096" containerID="d55c97ed3c41f0e468f9ad663a39b7882f411f90e2839331726692d718cabab2" exitCode=0 Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.143170 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6nmxg" event={"ID":"279f32be-baa9-4cf7-a540-f6f8552c2096","Type":"ContainerDied","Data":"d55c97ed3c41f0e468f9ad663a39b7882f411f90e2839331726692d718cabab2"} Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.143281 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6nmxg" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.143306 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6nmxg" event={"ID":"279f32be-baa9-4cf7-a540-f6f8552c2096","Type":"ContainerDied","Data":"c57818078a5b12218bb8ad3b425ca7bbe4a25f4791f7811e8930737480c13ee4"} Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.143346 4814 scope.go:117] "RemoveContainer" containerID="d55c97ed3c41f0e468f9ad663a39b7882f411f90e2839331726692d718cabab2" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.145996 4814 generic.go:334] "Generic (PLEG): container finished" podID="9e9625fe-2036-4308-a07f-0548705c91dc" containerID="fb7d6b6f669010deeb79746a3a9322ba7fe2ccf0a193dc249340379d744a69d2" exitCode=0 Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.146077 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n275h" event={"ID":"9e9625fe-2036-4308-a07f-0548705c91dc","Type":"ContainerDied","Data":"fb7d6b6f669010deeb79746a3a9322ba7fe2ccf0a193dc249340379d744a69d2"} Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.146127 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n275h" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.146124 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n275h" event={"ID":"9e9625fe-2036-4308-a07f-0548705c91dc","Type":"ContainerDied","Data":"e17f7ac292a2c9554e25f86b9e190dbfc6fd3bbd9e08ac18a4fc6ec758b52de5"} Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.190899 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n275h"] Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.194735 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-n275h"] Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.228176 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/279f32be-baa9-4cf7-a540-f6f8552c2096-catalog-content\") pod \"279f32be-baa9-4cf7-a540-f6f8552c2096\" (UID: \"279f32be-baa9-4cf7-a540-f6f8552c2096\") " Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.228470 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2zc4\" (UniqueName: \"kubernetes.io/projected/279f32be-baa9-4cf7-a540-f6f8552c2096-kube-api-access-s2zc4\") pod \"279f32be-baa9-4cf7-a540-f6f8552c2096\" (UID: \"279f32be-baa9-4cf7-a540-f6f8552c2096\") " Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.228599 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/279f32be-baa9-4cf7-a540-f6f8552c2096-utilities\") pod \"279f32be-baa9-4cf7-a540-f6f8552c2096\" (UID: \"279f32be-baa9-4cf7-a540-f6f8552c2096\") " Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.231114 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/279f32be-baa9-4cf7-a540-f6f8552c2096-utilities" (OuterVolumeSpecName: "utilities") pod "279f32be-baa9-4cf7-a540-f6f8552c2096" (UID: "279f32be-baa9-4cf7-a540-f6f8552c2096"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.235085 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/279f32be-baa9-4cf7-a540-f6f8552c2096-kube-api-access-s2zc4" (OuterVolumeSpecName: "kube-api-access-s2zc4") pod "279f32be-baa9-4cf7-a540-f6f8552c2096" (UID: "279f32be-baa9-4cf7-a540-f6f8552c2096"). InnerVolumeSpecName "kube-api-access-s2zc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.330185 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2zc4\" (UniqueName: \"kubernetes.io/projected/279f32be-baa9-4cf7-a540-f6f8552c2096-kube-api-access-s2zc4\") on node \"crc\" DevicePath \"\"" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.330226 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/279f32be-baa9-4cf7-a540-f6f8552c2096-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.372949 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/279f32be-baa9-4cf7-a540-f6f8552c2096-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "279f32be-baa9-4cf7-a540-f6f8552c2096" (UID: "279f32be-baa9-4cf7-a540-f6f8552c2096"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.431715 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/279f32be-baa9-4cf7-a540-f6f8552c2096-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.487146 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6nmxg"] Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.492551 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6nmxg"] Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.651074 4814 scope.go:117] "RemoveContainer" containerID="939a5b115173f39ef6bcad9a2ce834bcc8e4e525ba89fd98afe40d9e16462527" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.707725 4814 scope.go:117] "RemoveContainer" containerID="cb17c9698a9b5e8811b9b09aa304872540f2c25e919a4d378598fd17ead2ef3e" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.735819 4814 scope.go:117] "RemoveContainer" containerID="d55c97ed3c41f0e468f9ad663a39b7882f411f90e2839331726692d718cabab2" Feb 27 16:37:41 crc kubenswrapper[4814]: E0227 16:37:41.736989 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d55c97ed3c41f0e468f9ad663a39b7882f411f90e2839331726692d718cabab2\": container with ID starting with d55c97ed3c41f0e468f9ad663a39b7882f411f90e2839331726692d718cabab2 not found: ID does not exist" containerID="d55c97ed3c41f0e468f9ad663a39b7882f411f90e2839331726692d718cabab2" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.737037 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d55c97ed3c41f0e468f9ad663a39b7882f411f90e2839331726692d718cabab2"} err="failed to get container status \"d55c97ed3c41f0e468f9ad663a39b7882f411f90e2839331726692d718cabab2\": rpc error: code = NotFound desc = could not find container \"d55c97ed3c41f0e468f9ad663a39b7882f411f90e2839331726692d718cabab2\": container with ID starting with d55c97ed3c41f0e468f9ad663a39b7882f411f90e2839331726692d718cabab2 not found: ID does not exist" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.737069 4814 scope.go:117] "RemoveContainer" containerID="939a5b115173f39ef6bcad9a2ce834bcc8e4e525ba89fd98afe40d9e16462527" Feb 27 16:37:41 crc kubenswrapper[4814]: E0227 16:37:41.737990 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"939a5b115173f39ef6bcad9a2ce834bcc8e4e525ba89fd98afe40d9e16462527\": container with ID starting with 939a5b115173f39ef6bcad9a2ce834bcc8e4e525ba89fd98afe40d9e16462527 not found: ID does not exist" containerID="939a5b115173f39ef6bcad9a2ce834bcc8e4e525ba89fd98afe40d9e16462527" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.738066 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"939a5b115173f39ef6bcad9a2ce834bcc8e4e525ba89fd98afe40d9e16462527"} err="failed to get container status \"939a5b115173f39ef6bcad9a2ce834bcc8e4e525ba89fd98afe40d9e16462527\": rpc error: code = NotFound desc = could not find container \"939a5b115173f39ef6bcad9a2ce834bcc8e4e525ba89fd98afe40d9e16462527\": container with ID starting with 939a5b115173f39ef6bcad9a2ce834bcc8e4e525ba89fd98afe40d9e16462527 not found: ID does not exist" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.738115 4814 scope.go:117] "RemoveContainer" containerID="cb17c9698a9b5e8811b9b09aa304872540f2c25e919a4d378598fd17ead2ef3e" Feb 27 16:37:41 crc kubenswrapper[4814]: E0227 16:37:41.738671 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb17c9698a9b5e8811b9b09aa304872540f2c25e919a4d378598fd17ead2ef3e\": container with ID starting with cb17c9698a9b5e8811b9b09aa304872540f2c25e919a4d378598fd17ead2ef3e not found: ID does not exist" containerID="cb17c9698a9b5e8811b9b09aa304872540f2c25e919a4d378598fd17ead2ef3e" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.738704 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb17c9698a9b5e8811b9b09aa304872540f2c25e919a4d378598fd17ead2ef3e"} err="failed to get container status \"cb17c9698a9b5e8811b9b09aa304872540f2c25e919a4d378598fd17ead2ef3e\": rpc error: code = NotFound desc = could not find container \"cb17c9698a9b5e8811b9b09aa304872540f2c25e919a4d378598fd17ead2ef3e\": container with ID starting with cb17c9698a9b5e8811b9b09aa304872540f2c25e919a4d378598fd17ead2ef3e not found: ID does not exist" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.738720 4814 scope.go:117] "RemoveContainer" containerID="fb7d6b6f669010deeb79746a3a9322ba7fe2ccf0a193dc249340379d744a69d2" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.765892 4814 scope.go:117] "RemoveContainer" containerID="927158e6a433a614bf7d0ecbe61d1dc774f5efe22f20e4e6356ce727d11a425d" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.792987 4814 scope.go:117] "RemoveContainer" containerID="6784a11cc69dcdc5babbb8587209880debe22e3cf9c61c2d69ebe26b82bae460" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.835802 4814 scope.go:117] "RemoveContainer" containerID="fb7d6b6f669010deeb79746a3a9322ba7fe2ccf0a193dc249340379d744a69d2" Feb 27 16:37:41 crc kubenswrapper[4814]: E0227 16:37:41.837387 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb7d6b6f669010deeb79746a3a9322ba7fe2ccf0a193dc249340379d744a69d2\": container with ID starting with fb7d6b6f669010deeb79746a3a9322ba7fe2ccf0a193dc249340379d744a69d2 not found: ID does not exist" containerID="fb7d6b6f669010deeb79746a3a9322ba7fe2ccf0a193dc249340379d744a69d2" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.837426 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb7d6b6f669010deeb79746a3a9322ba7fe2ccf0a193dc249340379d744a69d2"} err="failed to get container status \"fb7d6b6f669010deeb79746a3a9322ba7fe2ccf0a193dc249340379d744a69d2\": rpc error: code = NotFound desc = could not find container \"fb7d6b6f669010deeb79746a3a9322ba7fe2ccf0a193dc249340379d744a69d2\": container with ID starting with fb7d6b6f669010deeb79746a3a9322ba7fe2ccf0a193dc249340379d744a69d2 not found: ID does not exist" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.837453 4814 scope.go:117] "RemoveContainer" containerID="927158e6a433a614bf7d0ecbe61d1dc774f5efe22f20e4e6356ce727d11a425d" Feb 27 16:37:41 crc kubenswrapper[4814]: E0227 16:37:41.837901 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"927158e6a433a614bf7d0ecbe61d1dc774f5efe22f20e4e6356ce727d11a425d\": container with ID starting with 927158e6a433a614bf7d0ecbe61d1dc774f5efe22f20e4e6356ce727d11a425d not found: ID does not exist" containerID="927158e6a433a614bf7d0ecbe61d1dc774f5efe22f20e4e6356ce727d11a425d" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.838057 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"927158e6a433a614bf7d0ecbe61d1dc774f5efe22f20e4e6356ce727d11a425d"} err="failed to get container status \"927158e6a433a614bf7d0ecbe61d1dc774f5efe22f20e4e6356ce727d11a425d\": rpc error: code = NotFound desc = could not find container \"927158e6a433a614bf7d0ecbe61d1dc774f5efe22f20e4e6356ce727d11a425d\": container with ID starting with 927158e6a433a614bf7d0ecbe61d1dc774f5efe22f20e4e6356ce727d11a425d not found: ID does not exist" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.838115 4814 scope.go:117] "RemoveContainer" containerID="6784a11cc69dcdc5babbb8587209880debe22e3cf9c61c2d69ebe26b82bae460" Feb 27 16:37:41 crc kubenswrapper[4814]: E0227 16:37:41.838712 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6784a11cc69dcdc5babbb8587209880debe22e3cf9c61c2d69ebe26b82bae460\": container with ID starting with 6784a11cc69dcdc5babbb8587209880debe22e3cf9c61c2d69ebe26b82bae460 not found: ID does not exist" containerID="6784a11cc69dcdc5babbb8587209880debe22e3cf9c61c2d69ebe26b82bae460" Feb 27 16:37:41 crc kubenswrapper[4814]: I0227 16:37:41.838748 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6784a11cc69dcdc5babbb8587209880debe22e3cf9c61c2d69ebe26b82bae460"} err="failed to get container status \"6784a11cc69dcdc5babbb8587209880debe22e3cf9c61c2d69ebe26b82bae460\": rpc error: code = NotFound desc = could not find container \"6784a11cc69dcdc5babbb8587209880debe22e3cf9c61c2d69ebe26b82bae460\": container with ID starting with 6784a11cc69dcdc5babbb8587209880debe22e3cf9c61c2d69ebe26b82bae460 not found: ID does not exist" Feb 27 16:37:42 crc kubenswrapper[4814]: I0227 16:37:42.157345 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-69594cc75-9xknj" event={"ID":"6df39380-a69c-4ef3-94ba-6a5ef31e7f50","Type":"ContainerStarted","Data":"9fd932b56ab584af94d8d267c774af8fe8f84a8d782288f8761d0d9d5ad24518"} Feb 27 16:37:42 crc kubenswrapper[4814]: I0227 16:37:42.183452 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-69594cc75-9xknj" podStartSLOduration=1.9374512419999999 podStartE2EDuration="8.183432827s" podCreationTimestamp="2026-02-27 16:37:34 +0000 UTC" firstStartedPulling="2026-02-27 16:37:35.466304669 +0000 UTC m=+867.918929499" lastFinishedPulling="2026-02-27 16:37:41.712286224 +0000 UTC m=+874.164911084" observedRunningTime="2026-02-27 16:37:42.179404676 +0000 UTC m=+874.632029536" watchObservedRunningTime="2026-02-27 16:37:42.183432827 +0000 UTC m=+874.636057667" Feb 27 16:37:42 crc kubenswrapper[4814]: I0227 16:37:42.498749 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="279f32be-baa9-4cf7-a540-f6f8552c2096" path="/var/lib/kubelet/pods/279f32be-baa9-4cf7-a540-f6f8552c2096/volumes" Feb 27 16:37:42 crc kubenswrapper[4814]: I0227 16:37:42.500455 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e9625fe-2036-4308-a07f-0548705c91dc" path="/var/lib/kubelet/pods/9e9625fe-2036-4308-a07f-0548705c91dc/volumes" Feb 27 16:37:45 crc kubenswrapper[4814]: I0227 16:37:45.577206 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-6j6pw" Feb 27 16:37:45 crc kubenswrapper[4814]: I0227 16:37:45.596454 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:45 crc kubenswrapper[4814]: I0227 16:37:45.597075 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:45 crc kubenswrapper[4814]: I0227 16:37:45.616091 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:46 crc kubenswrapper[4814]: I0227 16:37:46.199408 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7fcd9679d-dtpk9" Feb 27 16:37:46 crc kubenswrapper[4814]: I0227 16:37:46.282218 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-7x2qp"] Feb 27 16:37:52 crc kubenswrapper[4814]: I0227 16:37:52.903335 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:37:52 crc kubenswrapper[4814]: I0227 16:37:52.904619 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:37:55 crc kubenswrapper[4814]: I0227 16:37:55.216622 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-786f45cff4-bbbjp" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.147445 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536838-6bsv2"] Feb 27 16:38:00 crc kubenswrapper[4814]: E0227 16:38:00.148097 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9625fe-2036-4308-a07f-0548705c91dc" containerName="extract-content" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.148118 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9625fe-2036-4308-a07f-0548705c91dc" containerName="extract-content" Feb 27 16:38:00 crc kubenswrapper[4814]: E0227 16:38:00.148133 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="279f32be-baa9-4cf7-a540-f6f8552c2096" containerName="extract-utilities" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.148146 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="279f32be-baa9-4cf7-a540-f6f8552c2096" containerName="extract-utilities" Feb 27 16:38:00 crc kubenswrapper[4814]: E0227 16:38:00.148172 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="279f32be-baa9-4cf7-a540-f6f8552c2096" containerName="registry-server" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.148186 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="279f32be-baa9-4cf7-a540-f6f8552c2096" containerName="registry-server" Feb 27 16:38:00 crc kubenswrapper[4814]: E0227 16:38:00.148210 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9625fe-2036-4308-a07f-0548705c91dc" containerName="registry-server" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.148223 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9625fe-2036-4308-a07f-0548705c91dc" containerName="registry-server" Feb 27 16:38:00 crc kubenswrapper[4814]: E0227 16:38:00.148247 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="279f32be-baa9-4cf7-a540-f6f8552c2096" containerName="extract-content" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.148284 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="279f32be-baa9-4cf7-a540-f6f8552c2096" containerName="extract-content" Feb 27 16:38:00 crc kubenswrapper[4814]: E0227 16:38:00.148311 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9625fe-2036-4308-a07f-0548705c91dc" containerName="extract-utilities" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.148324 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9625fe-2036-4308-a07f-0548705c91dc" containerName="extract-utilities" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.148494 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9625fe-2036-4308-a07f-0548705c91dc" containerName="registry-server" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.148510 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="279f32be-baa9-4cf7-a540-f6f8552c2096" containerName="registry-server" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.149106 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536838-6bsv2" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.151728 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.152513 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.152836 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.157722 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536838-6bsv2"] Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.269889 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l6fm\" (UniqueName: \"kubernetes.io/projected/aa4535b0-2dac-45e8-9a9a-3ea58bb75658-kube-api-access-6l6fm\") pod \"auto-csr-approver-29536838-6bsv2\" (UID: \"aa4535b0-2dac-45e8-9a9a-3ea58bb75658\") " pod="openshift-infra/auto-csr-approver-29536838-6bsv2" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.371202 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l6fm\" (UniqueName: \"kubernetes.io/projected/aa4535b0-2dac-45e8-9a9a-3ea58bb75658-kube-api-access-6l6fm\") pod \"auto-csr-approver-29536838-6bsv2\" (UID: \"aa4535b0-2dac-45e8-9a9a-3ea58bb75658\") " pod="openshift-infra/auto-csr-approver-29536838-6bsv2" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.403619 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l6fm\" (UniqueName: \"kubernetes.io/projected/aa4535b0-2dac-45e8-9a9a-3ea58bb75658-kube-api-access-6l6fm\") pod \"auto-csr-approver-29536838-6bsv2\" (UID: \"aa4535b0-2dac-45e8-9a9a-3ea58bb75658\") " pod="openshift-infra/auto-csr-approver-29536838-6bsv2" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.485106 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536838-6bsv2" Feb 27 16:38:00 crc kubenswrapper[4814]: I0227 16:38:00.724657 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536838-6bsv2"] Feb 27 16:38:00 crc kubenswrapper[4814]: W0227 16:38:00.733405 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa4535b0_2dac_45e8_9a9a_3ea58bb75658.slice/crio-d01e25977768c5691ef0fa5e522b7eb4e8505d5e3e435a246d4d39d7cccf0c81 WatchSource:0}: Error finding container d01e25977768c5691ef0fa5e522b7eb4e8505d5e3e435a246d4d39d7cccf0c81: Status 404 returned error can't find the container with id d01e25977768c5691ef0fa5e522b7eb4e8505d5e3e435a246d4d39d7cccf0c81 Feb 27 16:38:01 crc kubenswrapper[4814]: I0227 16:38:01.358929 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536838-6bsv2" event={"ID":"aa4535b0-2dac-45e8-9a9a-3ea58bb75658","Type":"ContainerStarted","Data":"d01e25977768c5691ef0fa5e522b7eb4e8505d5e3e435a246d4d39d7cccf0c81"} Feb 27 16:38:02 crc kubenswrapper[4814]: I0227 16:38:02.370454 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536838-6bsv2" event={"ID":"aa4535b0-2dac-45e8-9a9a-3ea58bb75658","Type":"ContainerStarted","Data":"d0a9141a1d433581ab43efa57803ac8b74c0ad746b4ca3cd3c3b8657264681da"} Feb 27 16:38:03 crc kubenswrapper[4814]: I0227 16:38:03.382701 4814 generic.go:334] "Generic (PLEG): container finished" podID="aa4535b0-2dac-45e8-9a9a-3ea58bb75658" containerID="d0a9141a1d433581ab43efa57803ac8b74c0ad746b4ca3cd3c3b8657264681da" exitCode=0 Feb 27 16:38:03 crc kubenswrapper[4814]: I0227 16:38:03.382764 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536838-6bsv2" event={"ID":"aa4535b0-2dac-45e8-9a9a-3ea58bb75658","Type":"ContainerDied","Data":"d0a9141a1d433581ab43efa57803ac8b74c0ad746b4ca3cd3c3b8657264681da"} Feb 27 16:38:04 crc kubenswrapper[4814]: I0227 16:38:04.753350 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536838-6bsv2" Feb 27 16:38:04 crc kubenswrapper[4814]: I0227 16:38:04.840152 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6l6fm\" (UniqueName: \"kubernetes.io/projected/aa4535b0-2dac-45e8-9a9a-3ea58bb75658-kube-api-access-6l6fm\") pod \"aa4535b0-2dac-45e8-9a9a-3ea58bb75658\" (UID: \"aa4535b0-2dac-45e8-9a9a-3ea58bb75658\") " Feb 27 16:38:04 crc kubenswrapper[4814]: I0227 16:38:04.848074 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa4535b0-2dac-45e8-9a9a-3ea58bb75658-kube-api-access-6l6fm" (OuterVolumeSpecName: "kube-api-access-6l6fm") pod "aa4535b0-2dac-45e8-9a9a-3ea58bb75658" (UID: "aa4535b0-2dac-45e8-9a9a-3ea58bb75658"). InnerVolumeSpecName "kube-api-access-6l6fm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:38:04 crc kubenswrapper[4814]: I0227 16:38:04.941894 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6l6fm\" (UniqueName: \"kubernetes.io/projected/aa4535b0-2dac-45e8-9a9a-3ea58bb75658-kube-api-access-6l6fm\") on node \"crc\" DevicePath \"\"" Feb 27 16:38:05 crc kubenswrapper[4814]: I0227 16:38:05.399407 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536838-6bsv2" event={"ID":"aa4535b0-2dac-45e8-9a9a-3ea58bb75658","Type":"ContainerDied","Data":"d01e25977768c5691ef0fa5e522b7eb4e8505d5e3e435a246d4d39d7cccf0c81"} Feb 27 16:38:05 crc kubenswrapper[4814]: I0227 16:38:05.399453 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d01e25977768c5691ef0fa5e522b7eb4e8505d5e3e435a246d4d39d7cccf0c81" Feb 27 16:38:05 crc kubenswrapper[4814]: I0227 16:38:05.399497 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536838-6bsv2" Feb 27 16:38:05 crc kubenswrapper[4814]: I0227 16:38:05.867507 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536832-blzjj"] Feb 27 16:38:05 crc kubenswrapper[4814]: I0227 16:38:05.874256 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536832-blzjj"] Feb 27 16:38:06 crc kubenswrapper[4814]: I0227 16:38:06.501188 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="509692fb-2332-42cc-a678-b447aa1ed437" path="/var/lib/kubelet/pods/509692fb-2332-42cc-a678-b447aa1ed437/volumes" Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.369963 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-7x2qp" podUID="f487b7f4-c24b-431e-8868-a0285c8ed7bc" containerName="console" containerID="cri-o://6c107d6ef30a6b7c63776b99a6e0b9b6c74689983e810164c431b4553ef4f089" gracePeriod=15 Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.809919 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7x2qp_f487b7f4-c24b-431e-8868-a0285c8ed7bc/console/0.log" Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.810026 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.833094 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-oauth-config\") pod \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.833177 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-serving-cert\") pod \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.833300 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-trusted-ca-bundle\") pod \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.833349 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-service-ca\") pod \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.833403 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-oauth-serving-cert\") pod \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.833490 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-config\") pod \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.834688 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "f487b7f4-c24b-431e-8868-a0285c8ed7bc" (UID: "f487b7f4-c24b-431e-8868-a0285c8ed7bc"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.834710 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "f487b7f4-c24b-431e-8868-a0285c8ed7bc" (UID: "f487b7f4-c24b-431e-8868-a0285c8ed7bc"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.834747 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-service-ca" (OuterVolumeSpecName: "service-ca") pod "f487b7f4-c24b-431e-8868-a0285c8ed7bc" (UID: "f487b7f4-c24b-431e-8868-a0285c8ed7bc"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.834798 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmknd\" (UniqueName: \"kubernetes.io/projected/f487b7f4-c24b-431e-8868-a0285c8ed7bc-kube-api-access-qmknd\") pod \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\" (UID: \"f487b7f4-c24b-431e-8868-a0285c8ed7bc\") " Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.835306 4814 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.835334 4814 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.835354 4814 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.835492 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-config" (OuterVolumeSpecName: "console-config") pod "f487b7f4-c24b-431e-8868-a0285c8ed7bc" (UID: "f487b7f4-c24b-431e-8868-a0285c8ed7bc"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.873342 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "f487b7f4-c24b-431e-8868-a0285c8ed7bc" (UID: "f487b7f4-c24b-431e-8868-a0285c8ed7bc"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.873493 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f487b7f4-c24b-431e-8868-a0285c8ed7bc-kube-api-access-qmknd" (OuterVolumeSpecName: "kube-api-access-qmknd") pod "f487b7f4-c24b-431e-8868-a0285c8ed7bc" (UID: "f487b7f4-c24b-431e-8868-a0285c8ed7bc"). InnerVolumeSpecName "kube-api-access-qmknd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.874318 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "f487b7f4-c24b-431e-8868-a0285c8ed7bc" (UID: "f487b7f4-c24b-431e-8868-a0285c8ed7bc"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.939215 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmknd\" (UniqueName: \"kubernetes.io/projected/f487b7f4-c24b-431e-8868-a0285c8ed7bc-kube-api-access-qmknd\") on node \"crc\" DevicePath \"\"" Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.939295 4814 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.939316 4814 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:38:11 crc kubenswrapper[4814]: I0227 16:38:11.939334 4814 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f487b7f4-c24b-431e-8868-a0285c8ed7bc-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 16:38:12 crc kubenswrapper[4814]: I0227 16:38:12.455168 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7x2qp_f487b7f4-c24b-431e-8868-a0285c8ed7bc/console/0.log" Feb 27 16:38:12 crc kubenswrapper[4814]: I0227 16:38:12.455249 4814 generic.go:334] "Generic (PLEG): container finished" podID="f487b7f4-c24b-431e-8868-a0285c8ed7bc" containerID="6c107d6ef30a6b7c63776b99a6e0b9b6c74689983e810164c431b4553ef4f089" exitCode=2 Feb 27 16:38:12 crc kubenswrapper[4814]: I0227 16:38:12.455322 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7x2qp" event={"ID":"f487b7f4-c24b-431e-8868-a0285c8ed7bc","Type":"ContainerDied","Data":"6c107d6ef30a6b7c63776b99a6e0b9b6c74689983e810164c431b4553ef4f089"} Feb 27 16:38:12 crc kubenswrapper[4814]: I0227 16:38:12.455391 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7x2qp" event={"ID":"f487b7f4-c24b-431e-8868-a0285c8ed7bc","Type":"ContainerDied","Data":"e461de5943467cdbf6f7248a16c5dfd1c11ff1400fc8140953caf4be6f92c322"} Feb 27 16:38:12 crc kubenswrapper[4814]: I0227 16:38:12.455420 4814 scope.go:117] "RemoveContainer" containerID="6c107d6ef30a6b7c63776b99a6e0b9b6c74689983e810164c431b4553ef4f089" Feb 27 16:38:12 crc kubenswrapper[4814]: I0227 16:38:12.455581 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7x2qp" Feb 27 16:38:12 crc kubenswrapper[4814]: I0227 16:38:12.487665 4814 scope.go:117] "RemoveContainer" containerID="6c107d6ef30a6b7c63776b99a6e0b9b6c74689983e810164c431b4553ef4f089" Feb 27 16:38:12 crc kubenswrapper[4814]: E0227 16:38:12.488116 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c107d6ef30a6b7c63776b99a6e0b9b6c74689983e810164c431b4553ef4f089\": container with ID starting with 6c107d6ef30a6b7c63776b99a6e0b9b6c74689983e810164c431b4553ef4f089 not found: ID does not exist" containerID="6c107d6ef30a6b7c63776b99a6e0b9b6c74689983e810164c431b4553ef4f089" Feb 27 16:38:12 crc kubenswrapper[4814]: I0227 16:38:12.488153 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c107d6ef30a6b7c63776b99a6e0b9b6c74689983e810164c431b4553ef4f089"} err="failed to get container status \"6c107d6ef30a6b7c63776b99a6e0b9b6c74689983e810164c431b4553ef4f089\": rpc error: code = NotFound desc = could not find container \"6c107d6ef30a6b7c63776b99a6e0b9b6c74689983e810164c431b4553ef4f089\": container with ID starting with 6c107d6ef30a6b7c63776b99a6e0b9b6c74689983e810164c431b4553ef4f089 not found: ID does not exist" Feb 27 16:38:12 crc kubenswrapper[4814]: I0227 16:38:12.504329 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-7x2qp"] Feb 27 16:38:12 crc kubenswrapper[4814]: I0227 16:38:12.509754 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-7x2qp"] Feb 27 16:38:12 crc kubenswrapper[4814]: I0227 16:38:12.715985 4814 patch_prober.go:28] interesting pod/console-f9d7485db-7x2qp container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 16:38:12 crc kubenswrapper[4814]: I0227 16:38:12.716488 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-f9d7485db-7x2qp" podUID="f487b7f4-c24b-431e-8868-a0285c8ed7bc" containerName="console" probeResult="failure" output="Get \"https://10.217.0.26:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.642180 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj"] Feb 27 16:38:13 crc kubenswrapper[4814]: E0227 16:38:13.642673 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f487b7f4-c24b-431e-8868-a0285c8ed7bc" containerName="console" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.642695 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f487b7f4-c24b-431e-8868-a0285c8ed7bc" containerName="console" Feb 27 16:38:13 crc kubenswrapper[4814]: E0227 16:38:13.642742 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa4535b0-2dac-45e8-9a9a-3ea58bb75658" containerName="oc" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.642757 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa4535b0-2dac-45e8-9a9a-3ea58bb75658" containerName="oc" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.642940 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa4535b0-2dac-45e8-9a9a-3ea58bb75658" containerName="oc" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.642964 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="f487b7f4-c24b-431e-8868-a0285c8ed7bc" containerName="console" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.644481 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.646881 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.652907 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj"] Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.664043 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b790a0b4-30f0-40fa-9c06-6c91abe332e0-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj\" (UID: \"b790a0b4-30f0-40fa-9c06-6c91abe332e0\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.664143 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b790a0b4-30f0-40fa-9c06-6c91abe332e0-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj\" (UID: \"b790a0b4-30f0-40fa-9c06-6c91abe332e0\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.664236 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4vvs\" (UniqueName: \"kubernetes.io/projected/b790a0b4-30f0-40fa-9c06-6c91abe332e0-kube-api-access-z4vvs\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj\" (UID: \"b790a0b4-30f0-40fa-9c06-6c91abe332e0\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.765779 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4vvs\" (UniqueName: \"kubernetes.io/projected/b790a0b4-30f0-40fa-9c06-6c91abe332e0-kube-api-access-z4vvs\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj\" (UID: \"b790a0b4-30f0-40fa-9c06-6c91abe332e0\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.765925 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b790a0b4-30f0-40fa-9c06-6c91abe332e0-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj\" (UID: \"b790a0b4-30f0-40fa-9c06-6c91abe332e0\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.766008 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b790a0b4-30f0-40fa-9c06-6c91abe332e0-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj\" (UID: \"b790a0b4-30f0-40fa-9c06-6c91abe332e0\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.766708 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b790a0b4-30f0-40fa-9c06-6c91abe332e0-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj\" (UID: \"b790a0b4-30f0-40fa-9c06-6c91abe332e0\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.766817 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b790a0b4-30f0-40fa-9c06-6c91abe332e0-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj\" (UID: \"b790a0b4-30f0-40fa-9c06-6c91abe332e0\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.798749 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4vvs\" (UniqueName: \"kubernetes.io/projected/b790a0b4-30f0-40fa-9c06-6c91abe332e0-kube-api-access-z4vvs\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj\" (UID: \"b790a0b4-30f0-40fa-9c06-6c91abe332e0\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" Feb 27 16:38:13 crc kubenswrapper[4814]: I0227 16:38:13.964593 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" Feb 27 16:38:14 crc kubenswrapper[4814]: I0227 16:38:14.219434 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj"] Feb 27 16:38:14 crc kubenswrapper[4814]: W0227 16:38:14.226575 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb790a0b4_30f0_40fa_9c06_6c91abe332e0.slice/crio-4608a6c8032de404ca66b8736c501cff262db082e47ac2de127afc14faee15a6 WatchSource:0}: Error finding container 4608a6c8032de404ca66b8736c501cff262db082e47ac2de127afc14faee15a6: Status 404 returned error can't find the container with id 4608a6c8032de404ca66b8736c501cff262db082e47ac2de127afc14faee15a6 Feb 27 16:38:14 crc kubenswrapper[4814]: I0227 16:38:14.472502 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" event={"ID":"b790a0b4-30f0-40fa-9c06-6c91abe332e0","Type":"ContainerStarted","Data":"838084614bd3260c834e8e83c686be56109b60c9a906817a466728c65b8d581d"} Feb 27 16:38:14 crc kubenswrapper[4814]: I0227 16:38:14.472928 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" event={"ID":"b790a0b4-30f0-40fa-9c06-6c91abe332e0","Type":"ContainerStarted","Data":"4608a6c8032de404ca66b8736c501cff262db082e47ac2de127afc14faee15a6"} Feb 27 16:38:14 crc kubenswrapper[4814]: I0227 16:38:14.515860 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f487b7f4-c24b-431e-8868-a0285c8ed7bc" path="/var/lib/kubelet/pods/f487b7f4-c24b-431e-8868-a0285c8ed7bc/volumes" Feb 27 16:38:15 crc kubenswrapper[4814]: I0227 16:38:15.482765 4814 generic.go:334] "Generic (PLEG): container finished" podID="b790a0b4-30f0-40fa-9c06-6c91abe332e0" containerID="838084614bd3260c834e8e83c686be56109b60c9a906817a466728c65b8d581d" exitCode=0 Feb 27 16:38:15 crc kubenswrapper[4814]: I0227 16:38:15.482825 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" event={"ID":"b790a0b4-30f0-40fa-9c06-6c91abe332e0","Type":"ContainerDied","Data":"838084614bd3260c834e8e83c686be56109b60c9a906817a466728c65b8d581d"} Feb 27 16:38:17 crc kubenswrapper[4814]: I0227 16:38:17.501951 4814 generic.go:334] "Generic (PLEG): container finished" podID="b790a0b4-30f0-40fa-9c06-6c91abe332e0" containerID="0d2f449abdeebea9a9ffffb58826b4943d4842c67c4de3d1b7d6a43cea3b59f9" exitCode=0 Feb 27 16:38:17 crc kubenswrapper[4814]: I0227 16:38:17.502377 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" event={"ID":"b790a0b4-30f0-40fa-9c06-6c91abe332e0","Type":"ContainerDied","Data":"0d2f449abdeebea9a9ffffb58826b4943d4842c67c4de3d1b7d6a43cea3b59f9"} Feb 27 16:38:18 crc kubenswrapper[4814]: I0227 16:38:18.514518 4814 generic.go:334] "Generic (PLEG): container finished" podID="b790a0b4-30f0-40fa-9c06-6c91abe332e0" containerID="15ff98a9a8471481603bab512b4404cfcae63e7fc4a6cbb220e90f93fb6e2263" exitCode=0 Feb 27 16:38:18 crc kubenswrapper[4814]: I0227 16:38:18.514864 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" event={"ID":"b790a0b4-30f0-40fa-9c06-6c91abe332e0","Type":"ContainerDied","Data":"15ff98a9a8471481603bab512b4404cfcae63e7fc4a6cbb220e90f93fb6e2263"} Feb 27 16:38:19 crc kubenswrapper[4814]: I0227 16:38:19.867974 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" Feb 27 16:38:19 crc kubenswrapper[4814]: I0227 16:38:19.951062 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b790a0b4-30f0-40fa-9c06-6c91abe332e0-util\") pod \"b790a0b4-30f0-40fa-9c06-6c91abe332e0\" (UID: \"b790a0b4-30f0-40fa-9c06-6c91abe332e0\") " Feb 27 16:38:20 crc kubenswrapper[4814]: I0227 16:38:20.052891 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b790a0b4-30f0-40fa-9c06-6c91abe332e0-bundle\") pod \"b790a0b4-30f0-40fa-9c06-6c91abe332e0\" (UID: \"b790a0b4-30f0-40fa-9c06-6c91abe332e0\") " Feb 27 16:38:20 crc kubenswrapper[4814]: I0227 16:38:20.053108 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4vvs\" (UniqueName: \"kubernetes.io/projected/b790a0b4-30f0-40fa-9c06-6c91abe332e0-kube-api-access-z4vvs\") pod \"b790a0b4-30f0-40fa-9c06-6c91abe332e0\" (UID: \"b790a0b4-30f0-40fa-9c06-6c91abe332e0\") " Feb 27 16:38:20 crc kubenswrapper[4814]: I0227 16:38:20.054778 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b790a0b4-30f0-40fa-9c06-6c91abe332e0-bundle" (OuterVolumeSpecName: "bundle") pod "b790a0b4-30f0-40fa-9c06-6c91abe332e0" (UID: "b790a0b4-30f0-40fa-9c06-6c91abe332e0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:38:20 crc kubenswrapper[4814]: I0227 16:38:20.062153 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b790a0b4-30f0-40fa-9c06-6c91abe332e0-kube-api-access-z4vvs" (OuterVolumeSpecName: "kube-api-access-z4vvs") pod "b790a0b4-30f0-40fa-9c06-6c91abe332e0" (UID: "b790a0b4-30f0-40fa-9c06-6c91abe332e0"). InnerVolumeSpecName "kube-api-access-z4vvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:38:20 crc kubenswrapper[4814]: I0227 16:38:20.155463 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4vvs\" (UniqueName: \"kubernetes.io/projected/b790a0b4-30f0-40fa-9c06-6c91abe332e0-kube-api-access-z4vvs\") on node \"crc\" DevicePath \"\"" Feb 27 16:38:20 crc kubenswrapper[4814]: I0227 16:38:20.155533 4814 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b790a0b4-30f0-40fa-9c06-6c91abe332e0-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:38:20 crc kubenswrapper[4814]: I0227 16:38:20.310054 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b790a0b4-30f0-40fa-9c06-6c91abe332e0-util" (OuterVolumeSpecName: "util") pod "b790a0b4-30f0-40fa-9c06-6c91abe332e0" (UID: "b790a0b4-30f0-40fa-9c06-6c91abe332e0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:38:20 crc kubenswrapper[4814]: I0227 16:38:20.358166 4814 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b790a0b4-30f0-40fa-9c06-6c91abe332e0-util\") on node \"crc\" DevicePath \"\"" Feb 27 16:38:20 crc kubenswrapper[4814]: I0227 16:38:20.530745 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" event={"ID":"b790a0b4-30f0-40fa-9c06-6c91abe332e0","Type":"ContainerDied","Data":"4608a6c8032de404ca66b8736c501cff262db082e47ac2de127afc14faee15a6"} Feb 27 16:38:20 crc kubenswrapper[4814]: I0227 16:38:20.530783 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4608a6c8032de404ca66b8736c501cff262db082e47ac2de127afc14faee15a6" Feb 27 16:38:20 crc kubenswrapper[4814]: I0227 16:38:20.530836 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj" Feb 27 16:38:22 crc kubenswrapper[4814]: I0227 16:38:22.903056 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:38:22 crc kubenswrapper[4814]: I0227 16:38:22.903423 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:38:26 crc kubenswrapper[4814]: I0227 16:38:26.241575 4814 scope.go:117] "RemoveContainer" containerID="64aab1a2f7b2095ddb3c933a98f950873b3fa8b4b3a3564427139a7ffa933e44" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.624006 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f"] Feb 27 16:38:31 crc kubenswrapper[4814]: E0227 16:38:31.624631 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b790a0b4-30f0-40fa-9c06-6c91abe332e0" containerName="util" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.624642 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="b790a0b4-30f0-40fa-9c06-6c91abe332e0" containerName="util" Feb 27 16:38:31 crc kubenswrapper[4814]: E0227 16:38:31.624661 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b790a0b4-30f0-40fa-9c06-6c91abe332e0" containerName="pull" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.624668 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="b790a0b4-30f0-40fa-9c06-6c91abe332e0" containerName="pull" Feb 27 16:38:31 crc kubenswrapper[4814]: E0227 16:38:31.624678 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b790a0b4-30f0-40fa-9c06-6c91abe332e0" containerName="extract" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.624684 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="b790a0b4-30f0-40fa-9c06-6c91abe332e0" containerName="extract" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.624768 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="b790a0b4-30f0-40fa-9c06-6c91abe332e0" containerName="extract" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.625151 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.627305 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.627599 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.627689 4814 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.627787 4814 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.628765 4814 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-rsfsp" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.637530 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9-webhook-cert\") pod \"metallb-operator-controller-manager-58945cb6fc-nth4f\" (UID: \"e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9\") " pod="metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.637659 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6rp2\" (UniqueName: \"kubernetes.io/projected/e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9-kube-api-access-v6rp2\") pod \"metallb-operator-controller-manager-58945cb6fc-nth4f\" (UID: \"e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9\") " pod="metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.637699 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9-apiservice-cert\") pod \"metallb-operator-controller-manager-58945cb6fc-nth4f\" (UID: \"e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9\") " pod="metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.648971 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f"] Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.738287 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6rp2\" (UniqueName: \"kubernetes.io/projected/e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9-kube-api-access-v6rp2\") pod \"metallb-operator-controller-manager-58945cb6fc-nth4f\" (UID: \"e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9\") " pod="metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.738649 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9-apiservice-cert\") pod \"metallb-operator-controller-manager-58945cb6fc-nth4f\" (UID: \"e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9\") " pod="metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.739500 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9-webhook-cert\") pod \"metallb-operator-controller-manager-58945cb6fc-nth4f\" (UID: \"e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9\") " pod="metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.744326 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9-apiservice-cert\") pod \"metallb-operator-controller-manager-58945cb6fc-nth4f\" (UID: \"e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9\") " pod="metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.744396 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9-webhook-cert\") pod \"metallb-operator-controller-manager-58945cb6fc-nth4f\" (UID: \"e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9\") " pod="metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.753484 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6rp2\" (UniqueName: \"kubernetes.io/projected/e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9-kube-api-access-v6rp2\") pod \"metallb-operator-controller-manager-58945cb6fc-nth4f\" (UID: \"e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9\") " pod="metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.912527 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv"] Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.913173 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.916569 4814 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.916803 4814 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-rqtwh" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.916988 4814 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.933166 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv"] Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.940185 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.942417 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/20a110bc-385c-4cb3-a3e6-ce3568069d69-apiservice-cert\") pod \"metallb-operator-webhook-server-5d5d4f645d-nwdxv\" (UID: \"20a110bc-385c-4cb3-a3e6-ce3568069d69\") " pod="metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.942449 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/20a110bc-385c-4cb3-a3e6-ce3568069d69-webhook-cert\") pod \"metallb-operator-webhook-server-5d5d4f645d-nwdxv\" (UID: \"20a110bc-385c-4cb3-a3e6-ce3568069d69\") " pod="metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv" Feb 27 16:38:31 crc kubenswrapper[4814]: I0227 16:38:31.942492 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l656z\" (UniqueName: \"kubernetes.io/projected/20a110bc-385c-4cb3-a3e6-ce3568069d69-kube-api-access-l656z\") pod \"metallb-operator-webhook-server-5d5d4f645d-nwdxv\" (UID: \"20a110bc-385c-4cb3-a3e6-ce3568069d69\") " pod="metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv" Feb 27 16:38:32 crc kubenswrapper[4814]: I0227 16:38:32.044098 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/20a110bc-385c-4cb3-a3e6-ce3568069d69-apiservice-cert\") pod \"metallb-operator-webhook-server-5d5d4f645d-nwdxv\" (UID: \"20a110bc-385c-4cb3-a3e6-ce3568069d69\") " pod="metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv" Feb 27 16:38:32 crc kubenswrapper[4814]: I0227 16:38:32.044137 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/20a110bc-385c-4cb3-a3e6-ce3568069d69-webhook-cert\") pod \"metallb-operator-webhook-server-5d5d4f645d-nwdxv\" (UID: \"20a110bc-385c-4cb3-a3e6-ce3568069d69\") " pod="metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv" Feb 27 16:38:32 crc kubenswrapper[4814]: I0227 16:38:32.044185 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l656z\" (UniqueName: \"kubernetes.io/projected/20a110bc-385c-4cb3-a3e6-ce3568069d69-kube-api-access-l656z\") pod \"metallb-operator-webhook-server-5d5d4f645d-nwdxv\" (UID: \"20a110bc-385c-4cb3-a3e6-ce3568069d69\") " pod="metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv" Feb 27 16:38:32 crc kubenswrapper[4814]: I0227 16:38:32.051962 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/20a110bc-385c-4cb3-a3e6-ce3568069d69-apiservice-cert\") pod \"metallb-operator-webhook-server-5d5d4f645d-nwdxv\" (UID: \"20a110bc-385c-4cb3-a3e6-ce3568069d69\") " pod="metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv" Feb 27 16:38:32 crc kubenswrapper[4814]: I0227 16:38:32.051973 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/20a110bc-385c-4cb3-a3e6-ce3568069d69-webhook-cert\") pod \"metallb-operator-webhook-server-5d5d4f645d-nwdxv\" (UID: \"20a110bc-385c-4cb3-a3e6-ce3568069d69\") " pod="metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv" Feb 27 16:38:32 crc kubenswrapper[4814]: I0227 16:38:32.061156 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l656z\" (UniqueName: \"kubernetes.io/projected/20a110bc-385c-4cb3-a3e6-ce3568069d69-kube-api-access-l656z\") pod \"metallb-operator-webhook-server-5d5d4f645d-nwdxv\" (UID: \"20a110bc-385c-4cb3-a3e6-ce3568069d69\") " pod="metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv" Feb 27 16:38:32 crc kubenswrapper[4814]: I0227 16:38:32.228377 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv" Feb 27 16:38:32 crc kubenswrapper[4814]: I0227 16:38:32.409577 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f"] Feb 27 16:38:32 crc kubenswrapper[4814]: I0227 16:38:32.609940 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f" event={"ID":"e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9","Type":"ContainerStarted","Data":"81978ee81f5b426f026149b2296a30d8730de90bbd99cb5cddcc1eb849ad7076"} Feb 27 16:38:32 crc kubenswrapper[4814]: I0227 16:38:32.684142 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv"] Feb 27 16:38:32 crc kubenswrapper[4814]: W0227 16:38:32.684970 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20a110bc_385c_4cb3_a3e6_ce3568069d69.slice/crio-778e27cb854ff96d5fbe75354ae62752d9b13807681c7eb74faf39cd50b568c7 WatchSource:0}: Error finding container 778e27cb854ff96d5fbe75354ae62752d9b13807681c7eb74faf39cd50b568c7: Status 404 returned error can't find the container with id 778e27cb854ff96d5fbe75354ae62752d9b13807681c7eb74faf39cd50b568c7 Feb 27 16:38:33 crc kubenswrapper[4814]: I0227 16:38:33.618351 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv" event={"ID":"20a110bc-385c-4cb3-a3e6-ce3568069d69","Type":"ContainerStarted","Data":"778e27cb854ff96d5fbe75354ae62752d9b13807681c7eb74faf39cd50b568c7"} Feb 27 16:38:35 crc kubenswrapper[4814]: I0227 16:38:35.632129 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f" event={"ID":"e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9","Type":"ContainerStarted","Data":"e1fefe7e7e80241b741fcbde50ab9c7f72505c4cc885bb14865711c8d1ae31e5"} Feb 27 16:38:35 crc kubenswrapper[4814]: I0227 16:38:35.632928 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f" Feb 27 16:38:35 crc kubenswrapper[4814]: I0227 16:38:35.674033 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f" podStartSLOduration=1.7491100899999998 podStartE2EDuration="4.674003697s" podCreationTimestamp="2026-02-27 16:38:31 +0000 UTC" firstStartedPulling="2026-02-27 16:38:32.422819901 +0000 UTC m=+924.875444731" lastFinishedPulling="2026-02-27 16:38:35.347713508 +0000 UTC m=+927.800338338" observedRunningTime="2026-02-27 16:38:35.668370923 +0000 UTC m=+928.120995753" watchObservedRunningTime="2026-02-27 16:38:35.674003697 +0000 UTC m=+928.126628537" Feb 27 16:38:38 crc kubenswrapper[4814]: I0227 16:38:38.661063 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv" event={"ID":"20a110bc-385c-4cb3-a3e6-ce3568069d69","Type":"ContainerStarted","Data":"27d508c214ec3e9e936cd5c702bde31fb2e2c580e2467ee8f485f7e325bc1063"} Feb 27 16:38:38 crc kubenswrapper[4814]: I0227 16:38:38.663101 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv" Feb 27 16:38:38 crc kubenswrapper[4814]: I0227 16:38:38.692673 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv" podStartSLOduration=2.605608681 podStartE2EDuration="7.692625s" podCreationTimestamp="2026-02-27 16:38:31 +0000 UTC" firstStartedPulling="2026-02-27 16:38:32.689504672 +0000 UTC m=+925.142129502" lastFinishedPulling="2026-02-27 16:38:37.776520971 +0000 UTC m=+930.229145821" observedRunningTime="2026-02-27 16:38:38.685931633 +0000 UTC m=+931.138556503" watchObservedRunningTime="2026-02-27 16:38:38.692625 +0000 UTC m=+931.145249880" Feb 27 16:38:52 crc kubenswrapper[4814]: I0227 16:38:52.239428 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5d5d4f645d-nwdxv" Feb 27 16:38:52 crc kubenswrapper[4814]: I0227 16:38:52.902688 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:38:52 crc kubenswrapper[4814]: I0227 16:38:52.903196 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:38:52 crc kubenswrapper[4814]: I0227 16:38:52.903298 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:38:52 crc kubenswrapper[4814]: I0227 16:38:52.904295 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7a221694cd83011f5779a509b7774635f372c6adcfadbeae289b463611af2c58"} pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 16:38:52 crc kubenswrapper[4814]: I0227 16:38:52.904415 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" containerID="cri-o://7a221694cd83011f5779a509b7774635f372c6adcfadbeae289b463611af2c58" gracePeriod=600 Feb 27 16:38:53 crc kubenswrapper[4814]: I0227 16:38:53.789877 4814 generic.go:334] "Generic (PLEG): container finished" podID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerID="7a221694cd83011f5779a509b7774635f372c6adcfadbeae289b463611af2c58" exitCode=0 Feb 27 16:38:53 crc kubenswrapper[4814]: I0227 16:38:53.789978 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerDied","Data":"7a221694cd83011f5779a509b7774635f372c6adcfadbeae289b463611af2c58"} Feb 27 16:38:53 crc kubenswrapper[4814]: I0227 16:38:53.791471 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"be2cf4df4208edbaae217431871265acc14c77fe4583a6ab07012775c0d5e5f8"} Feb 27 16:38:53 crc kubenswrapper[4814]: I0227 16:38:53.791523 4814 scope.go:117] "RemoveContainer" containerID="5179547dd473be027ccf5a04ad384d8f4660f7e3ba2b8851cd7be0f851fdbb8a" Feb 27 16:39:11 crc kubenswrapper[4814]: I0227 16:39:11.944834 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-58945cb6fc-nth4f" Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.867678 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-lt84m"] Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.868836 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-lt84m" Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.874828 4814 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.874999 4814 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-4nwcw" Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.888984 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-5r8hf"] Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.891994 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.892687 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-lt84m"] Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.894145 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.894372 4814 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.935873 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-v59vj"] Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.937187 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-v59vj" Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.941312 4814 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-j4rfj" Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.941861 4814 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.942018 4814 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.942522 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.959003 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-86ddb6bd46-5lqjv"] Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.962229 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-86ddb6bd46-5lqjv" Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.965340 4814 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 27 16:39:12 crc kubenswrapper[4814]: I0227 16:39:12.974748 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-86ddb6bd46-5lqjv"] Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.066642 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/44587c86-594d-410c-939c-f3a1426ae15a-metallb-excludel2\") pod \"speaker-v59vj\" (UID: \"44587c86-594d-410c-939c-f3a1426ae15a\") " pod="metallb-system/speaker-v59vj" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.066718 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hdgt\" (UniqueName: \"kubernetes.io/projected/c484c9ee-2da0-4ec8-adf2-5ded9c667133-kube-api-access-6hdgt\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.066759 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a197efea-0275-49bc-acd5-b043737cd1a8-metrics-certs\") pod \"controller-86ddb6bd46-5lqjv\" (UID: \"a197efea-0275-49bc-acd5-b043737cd1a8\") " pod="metallb-system/controller-86ddb6bd46-5lqjv" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.066787 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c484c9ee-2da0-4ec8-adf2-5ded9c667133-frr-sockets\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.066812 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99q4j\" (UniqueName: \"kubernetes.io/projected/bbc12f0a-9041-484e-9216-9a618b63257b-kube-api-access-99q4j\") pod \"frr-k8s-webhook-server-7f989f654f-lt84m\" (UID: \"bbc12f0a-9041-484e-9216-9a618b63257b\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-lt84m" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.066833 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c484c9ee-2da0-4ec8-adf2-5ded9c667133-reloader\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.066864 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm4qb\" (UniqueName: \"kubernetes.io/projected/a197efea-0275-49bc-acd5-b043737cd1a8-kube-api-access-tm4qb\") pod \"controller-86ddb6bd46-5lqjv\" (UID: \"a197efea-0275-49bc-acd5-b043737cd1a8\") " pod="metallb-system/controller-86ddb6bd46-5lqjv" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.066917 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a197efea-0275-49bc-acd5-b043737cd1a8-cert\") pod \"controller-86ddb6bd46-5lqjv\" (UID: \"a197efea-0275-49bc-acd5-b043737cd1a8\") " pod="metallb-system/controller-86ddb6bd46-5lqjv" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.066943 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bbc12f0a-9041-484e-9216-9a618b63257b-cert\") pod \"frr-k8s-webhook-server-7f989f654f-lt84m\" (UID: \"bbc12f0a-9041-484e-9216-9a618b63257b\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-lt84m" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.066969 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c484c9ee-2da0-4ec8-adf2-5ded9c667133-frr-startup\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.066997 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c484c9ee-2da0-4ec8-adf2-5ded9c667133-metrics-certs\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.067029 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhvmh\" (UniqueName: \"kubernetes.io/projected/44587c86-594d-410c-939c-f3a1426ae15a-kube-api-access-mhvmh\") pod \"speaker-v59vj\" (UID: \"44587c86-594d-410c-939c-f3a1426ae15a\") " pod="metallb-system/speaker-v59vj" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.067057 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c484c9ee-2da0-4ec8-adf2-5ded9c667133-frr-conf\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.067087 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c484c9ee-2da0-4ec8-adf2-5ded9c667133-metrics\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.067116 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/44587c86-594d-410c-939c-f3a1426ae15a-memberlist\") pod \"speaker-v59vj\" (UID: \"44587c86-594d-410c-939c-f3a1426ae15a\") " pod="metallb-system/speaker-v59vj" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.067147 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/44587c86-594d-410c-939c-f3a1426ae15a-metrics-certs\") pod \"speaker-v59vj\" (UID: \"44587c86-594d-410c-939c-f3a1426ae15a\") " pod="metallb-system/speaker-v59vj" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.168524 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhvmh\" (UniqueName: \"kubernetes.io/projected/44587c86-594d-410c-939c-f3a1426ae15a-kube-api-access-mhvmh\") pod \"speaker-v59vj\" (UID: \"44587c86-594d-410c-939c-f3a1426ae15a\") " pod="metallb-system/speaker-v59vj" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.169100 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c484c9ee-2da0-4ec8-adf2-5ded9c667133-frr-conf\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.169721 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c484c9ee-2da0-4ec8-adf2-5ded9c667133-frr-conf\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.169803 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c484c9ee-2da0-4ec8-adf2-5ded9c667133-metrics\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.170125 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c484c9ee-2da0-4ec8-adf2-5ded9c667133-metrics\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.170173 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/44587c86-594d-410c-939c-f3a1426ae15a-memberlist\") pod \"speaker-v59vj\" (UID: \"44587c86-594d-410c-939c-f3a1426ae15a\") " pod="metallb-system/speaker-v59vj" Feb 27 16:39:13 crc kubenswrapper[4814]: E0227 16:39:13.170197 4814 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 27 16:39:13 crc kubenswrapper[4814]: E0227 16:39:13.170334 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/44587c86-594d-410c-939c-f3a1426ae15a-memberlist podName:44587c86-594d-410c-939c-f3a1426ae15a nodeName:}" failed. No retries permitted until 2026-02-27 16:39:13.670302022 +0000 UTC m=+966.122926872 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/44587c86-594d-410c-939c-f3a1426ae15a-memberlist") pod "speaker-v59vj" (UID: "44587c86-594d-410c-939c-f3a1426ae15a") : secret "metallb-memberlist" not found Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.170243 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/44587c86-594d-410c-939c-f3a1426ae15a-metrics-certs\") pod \"speaker-v59vj\" (UID: \"44587c86-594d-410c-939c-f3a1426ae15a\") " pod="metallb-system/speaker-v59vj" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.171740 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/44587c86-594d-410c-939c-f3a1426ae15a-metallb-excludel2\") pod \"speaker-v59vj\" (UID: \"44587c86-594d-410c-939c-f3a1426ae15a\") " pod="metallb-system/speaker-v59vj" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.171771 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hdgt\" (UniqueName: \"kubernetes.io/projected/c484c9ee-2da0-4ec8-adf2-5ded9c667133-kube-api-access-6hdgt\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.171808 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a197efea-0275-49bc-acd5-b043737cd1a8-metrics-certs\") pod \"controller-86ddb6bd46-5lqjv\" (UID: \"a197efea-0275-49bc-acd5-b043737cd1a8\") " pod="metallb-system/controller-86ddb6bd46-5lqjv" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.171868 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c484c9ee-2da0-4ec8-adf2-5ded9c667133-frr-sockets\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: E0227 16:39:13.171991 4814 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Feb 27 16:39:13 crc kubenswrapper[4814]: E0227 16:39:13.172067 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a197efea-0275-49bc-acd5-b043737cd1a8-metrics-certs podName:a197efea-0275-49bc-acd5-b043737cd1a8 nodeName:}" failed. No retries permitted until 2026-02-27 16:39:13.672041046 +0000 UTC m=+966.124665896 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a197efea-0275-49bc-acd5-b043737cd1a8-metrics-certs") pod "controller-86ddb6bd46-5lqjv" (UID: "a197efea-0275-49bc-acd5-b043737cd1a8") : secret "controller-certs-secret" not found Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.172099 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99q4j\" (UniqueName: \"kubernetes.io/projected/bbc12f0a-9041-484e-9216-9a618b63257b-kube-api-access-99q4j\") pod \"frr-k8s-webhook-server-7f989f654f-lt84m\" (UID: \"bbc12f0a-9041-484e-9216-9a618b63257b\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-lt84m" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.172127 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c484c9ee-2da0-4ec8-adf2-5ded9c667133-reloader\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.172461 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm4qb\" (UniqueName: \"kubernetes.io/projected/a197efea-0275-49bc-acd5-b043737cd1a8-kube-api-access-tm4qb\") pod \"controller-86ddb6bd46-5lqjv\" (UID: \"a197efea-0275-49bc-acd5-b043737cd1a8\") " pod="metallb-system/controller-86ddb6bd46-5lqjv" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.172639 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c484c9ee-2da0-4ec8-adf2-5ded9c667133-reloader\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.172719 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a197efea-0275-49bc-acd5-b043737cd1a8-cert\") pod \"controller-86ddb6bd46-5lqjv\" (UID: \"a197efea-0275-49bc-acd5-b043737cd1a8\") " pod="metallb-system/controller-86ddb6bd46-5lqjv" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.172753 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bbc12f0a-9041-484e-9216-9a618b63257b-cert\") pod \"frr-k8s-webhook-server-7f989f654f-lt84m\" (UID: \"bbc12f0a-9041-484e-9216-9a618b63257b\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-lt84m" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.172792 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c484c9ee-2da0-4ec8-adf2-5ded9c667133-frr-startup\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.172721 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c484c9ee-2da0-4ec8-adf2-5ded9c667133-frr-sockets\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.172828 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c484c9ee-2da0-4ec8-adf2-5ded9c667133-metrics-certs\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.173429 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/44587c86-594d-410c-939c-f3a1426ae15a-metallb-excludel2\") pod \"speaker-v59vj\" (UID: \"44587c86-594d-410c-939c-f3a1426ae15a\") " pod="metallb-system/speaker-v59vj" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.173815 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c484c9ee-2da0-4ec8-adf2-5ded9c667133-frr-startup\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.174248 4814 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.178766 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c484c9ee-2da0-4ec8-adf2-5ded9c667133-metrics-certs\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.178802 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/44587c86-594d-410c-939c-f3a1426ae15a-metrics-certs\") pod \"speaker-v59vj\" (UID: \"44587c86-594d-410c-939c-f3a1426ae15a\") " pod="metallb-system/speaker-v59vj" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.179521 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bbc12f0a-9041-484e-9216-9a618b63257b-cert\") pod \"frr-k8s-webhook-server-7f989f654f-lt84m\" (UID: \"bbc12f0a-9041-484e-9216-9a618b63257b\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-lt84m" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.185347 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a197efea-0275-49bc-acd5-b043737cd1a8-cert\") pod \"controller-86ddb6bd46-5lqjv\" (UID: \"a197efea-0275-49bc-acd5-b043737cd1a8\") " pod="metallb-system/controller-86ddb6bd46-5lqjv" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.188206 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99q4j\" (UniqueName: \"kubernetes.io/projected/bbc12f0a-9041-484e-9216-9a618b63257b-kube-api-access-99q4j\") pod \"frr-k8s-webhook-server-7f989f654f-lt84m\" (UID: \"bbc12f0a-9041-484e-9216-9a618b63257b\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-lt84m" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.193286 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm4qb\" (UniqueName: \"kubernetes.io/projected/a197efea-0275-49bc-acd5-b043737cd1a8-kube-api-access-tm4qb\") pod \"controller-86ddb6bd46-5lqjv\" (UID: \"a197efea-0275-49bc-acd5-b043737cd1a8\") " pod="metallb-system/controller-86ddb6bd46-5lqjv" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.195876 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hdgt\" (UniqueName: \"kubernetes.io/projected/c484c9ee-2da0-4ec8-adf2-5ded9c667133-kube-api-access-6hdgt\") pod \"frr-k8s-5r8hf\" (UID: \"c484c9ee-2da0-4ec8-adf2-5ded9c667133\") " pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.204068 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhvmh\" (UniqueName: \"kubernetes.io/projected/44587c86-594d-410c-939c-f3a1426ae15a-kube-api-access-mhvmh\") pod \"speaker-v59vj\" (UID: \"44587c86-594d-410c-939c-f3a1426ae15a\") " pod="metallb-system/speaker-v59vj" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.239440 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-lt84m" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.251652 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.687877 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/44587c86-594d-410c-939c-f3a1426ae15a-memberlist\") pod \"speaker-v59vj\" (UID: \"44587c86-594d-410c-939c-f3a1426ae15a\") " pod="metallb-system/speaker-v59vj" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.687990 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a197efea-0275-49bc-acd5-b043737cd1a8-metrics-certs\") pod \"controller-86ddb6bd46-5lqjv\" (UID: \"a197efea-0275-49bc-acd5-b043737cd1a8\") " pod="metallb-system/controller-86ddb6bd46-5lqjv" Feb 27 16:39:13 crc kubenswrapper[4814]: E0227 16:39:13.688141 4814 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 27 16:39:13 crc kubenswrapper[4814]: E0227 16:39:13.688316 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/44587c86-594d-410c-939c-f3a1426ae15a-memberlist podName:44587c86-594d-410c-939c-f3a1426ae15a nodeName:}" failed. No retries permitted until 2026-02-27 16:39:14.688288376 +0000 UTC m=+967.140913236 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/44587c86-594d-410c-939c-f3a1426ae15a-memberlist") pod "speaker-v59vj" (UID: "44587c86-594d-410c-939c-f3a1426ae15a") : secret "metallb-memberlist" not found Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.695825 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a197efea-0275-49bc-acd5-b043737cd1a8-metrics-certs\") pod \"controller-86ddb6bd46-5lqjv\" (UID: \"a197efea-0275-49bc-acd5-b043737cd1a8\") " pod="metallb-system/controller-86ddb6bd46-5lqjv" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.771302 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-lt84m"] Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.897678 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-86ddb6bd46-5lqjv" Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.951111 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-lt84m" event={"ID":"bbc12f0a-9041-484e-9216-9a618b63257b","Type":"ContainerStarted","Data":"1e6061532712b40789d9436032e18d9bc43ac8454c5e34174f1991ae59709dbb"} Feb 27 16:39:13 crc kubenswrapper[4814]: I0227 16:39:13.953205 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r8hf" event={"ID":"c484c9ee-2da0-4ec8-adf2-5ded9c667133","Type":"ContainerStarted","Data":"5403c2ca118b8775741ac84191c91ff1491f78f03aaf1d8a061adc20bdcddaa4"} Feb 27 16:39:14 crc kubenswrapper[4814]: I0227 16:39:14.238359 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-86ddb6bd46-5lqjv"] Feb 27 16:39:14 crc kubenswrapper[4814]: W0227 16:39:14.243879 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda197efea_0275_49bc_acd5_b043737cd1a8.slice/crio-65789003459ce2b0634ffa9ae76d50db6169dfb7b14ecfbf1932b297692950dd WatchSource:0}: Error finding container 65789003459ce2b0634ffa9ae76d50db6169dfb7b14ecfbf1932b297692950dd: Status 404 returned error can't find the container with id 65789003459ce2b0634ffa9ae76d50db6169dfb7b14ecfbf1932b297692950dd Feb 27 16:39:14 crc kubenswrapper[4814]: I0227 16:39:14.717758 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/44587c86-594d-410c-939c-f3a1426ae15a-memberlist\") pod \"speaker-v59vj\" (UID: \"44587c86-594d-410c-939c-f3a1426ae15a\") " pod="metallb-system/speaker-v59vj" Feb 27 16:39:14 crc kubenswrapper[4814]: I0227 16:39:14.723634 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/44587c86-594d-410c-939c-f3a1426ae15a-memberlist\") pod \"speaker-v59vj\" (UID: \"44587c86-594d-410c-939c-f3a1426ae15a\") " pod="metallb-system/speaker-v59vj" Feb 27 16:39:14 crc kubenswrapper[4814]: I0227 16:39:14.787009 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-v59vj" Feb 27 16:39:14 crc kubenswrapper[4814]: I0227 16:39:14.964703 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-5lqjv" event={"ID":"a197efea-0275-49bc-acd5-b043737cd1a8","Type":"ContainerStarted","Data":"440049487e120c1111e38875f05249ca64c1f8986896d48e2081be7fd67760be"} Feb 27 16:39:14 crc kubenswrapper[4814]: I0227 16:39:14.964766 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-5lqjv" event={"ID":"a197efea-0275-49bc-acd5-b043737cd1a8","Type":"ContainerStarted","Data":"b11344874e19f463d3bc8699eae97bb740f932988f0843bda58b0d1803a9d605"} Feb 27 16:39:14 crc kubenswrapper[4814]: I0227 16:39:14.964782 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-5lqjv" event={"ID":"a197efea-0275-49bc-acd5-b043737cd1a8","Type":"ContainerStarted","Data":"65789003459ce2b0634ffa9ae76d50db6169dfb7b14ecfbf1932b297692950dd"} Feb 27 16:39:14 crc kubenswrapper[4814]: I0227 16:39:14.964827 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-86ddb6bd46-5lqjv" Feb 27 16:39:14 crc kubenswrapper[4814]: I0227 16:39:14.969099 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-v59vj" event={"ID":"44587c86-594d-410c-939c-f3a1426ae15a","Type":"ContainerStarted","Data":"31a83338eacbe73c817386ef8b7828fa9294a135dfd29f4a2a9b225625311151"} Feb 27 16:39:14 crc kubenswrapper[4814]: I0227 16:39:14.987144 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-86ddb6bd46-5lqjv" podStartSLOduration=2.987125776 podStartE2EDuration="2.987125776s" podCreationTimestamp="2026-02-27 16:39:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:39:14.983920616 +0000 UTC m=+967.436545456" watchObservedRunningTime="2026-02-27 16:39:14.987125776 +0000 UTC m=+967.439750606" Feb 27 16:39:15 crc kubenswrapper[4814]: I0227 16:39:15.994873 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-v59vj" event={"ID":"44587c86-594d-410c-939c-f3a1426ae15a","Type":"ContainerStarted","Data":"432b78d96994f764cb7d0d1e91854e4e8fddfcb983ae215fc5538f23602fb4e8"} Feb 27 16:39:15 crc kubenswrapper[4814]: I0227 16:39:15.995208 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-v59vj" event={"ID":"44587c86-594d-410c-939c-f3a1426ae15a","Type":"ContainerStarted","Data":"d57a9fb9f60ebba4cee665e62241732de05ca3caa775a40abe8c56ac99400f20"} Feb 27 16:39:15 crc kubenswrapper[4814]: I0227 16:39:15.995226 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-v59vj" Feb 27 16:39:16 crc kubenswrapper[4814]: I0227 16:39:16.019961 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-v59vj" podStartSLOduration=4.019940145 podStartE2EDuration="4.019940145s" podCreationTimestamp="2026-02-27 16:39:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:39:16.017022975 +0000 UTC m=+968.469647815" watchObservedRunningTime="2026-02-27 16:39:16.019940145 +0000 UTC m=+968.472564975" Feb 27 16:39:22 crc kubenswrapper[4814]: I0227 16:39:22.053153 4814 generic.go:334] "Generic (PLEG): container finished" podID="c484c9ee-2da0-4ec8-adf2-5ded9c667133" containerID="ac2fcd5cba9736e9acd49737d145354584acb3f475593c005fbfaf368c82958c" exitCode=0 Feb 27 16:39:22 crc kubenswrapper[4814]: I0227 16:39:22.053232 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r8hf" event={"ID":"c484c9ee-2da0-4ec8-adf2-5ded9c667133","Type":"ContainerDied","Data":"ac2fcd5cba9736e9acd49737d145354584acb3f475593c005fbfaf368c82958c"} Feb 27 16:39:22 crc kubenswrapper[4814]: I0227 16:39:22.057201 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-lt84m" event={"ID":"bbc12f0a-9041-484e-9216-9a618b63257b","Type":"ContainerStarted","Data":"66796ea37f0895c550b5f458ab5fb4123f68a2da221280c0428483957d0ae2cd"} Feb 27 16:39:22 crc kubenswrapper[4814]: I0227 16:39:22.125471 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-lt84m" podStartSLOduration=2.576944104 podStartE2EDuration="10.125446009s" podCreationTimestamp="2026-02-27 16:39:12 +0000 UTC" firstStartedPulling="2026-02-27 16:39:13.788043039 +0000 UTC m=+966.240667879" lastFinishedPulling="2026-02-27 16:39:21.336544944 +0000 UTC m=+973.789169784" observedRunningTime="2026-02-27 16:39:22.121659782 +0000 UTC m=+974.574284652" watchObservedRunningTime="2026-02-27 16:39:22.125446009 +0000 UTC m=+974.578070849" Feb 27 16:39:23 crc kubenswrapper[4814]: I0227 16:39:23.067432 4814 generic.go:334] "Generic (PLEG): container finished" podID="c484c9ee-2da0-4ec8-adf2-5ded9c667133" containerID="c435dfdfab2042a48508f8526ebb4a291971efcaec9f4eb25ff980c2c2ffe57f" exitCode=0 Feb 27 16:39:23 crc kubenswrapper[4814]: I0227 16:39:23.067519 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r8hf" event={"ID":"c484c9ee-2da0-4ec8-adf2-5ded9c667133","Type":"ContainerDied","Data":"c435dfdfab2042a48508f8526ebb4a291971efcaec9f4eb25ff980c2c2ffe57f"} Feb 27 16:39:23 crc kubenswrapper[4814]: I0227 16:39:23.067685 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-lt84m" Feb 27 16:39:24 crc kubenswrapper[4814]: I0227 16:39:24.075316 4814 generic.go:334] "Generic (PLEG): container finished" podID="c484c9ee-2da0-4ec8-adf2-5ded9c667133" containerID="69c1c58087c09cdd4fc021b149684bd0c2e17c8d87677c6e8cdd043d99b6ee2c" exitCode=0 Feb 27 16:39:24 crc kubenswrapper[4814]: I0227 16:39:24.075434 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r8hf" event={"ID":"c484c9ee-2da0-4ec8-adf2-5ded9c667133","Type":"ContainerDied","Data":"69c1c58087c09cdd4fc021b149684bd0c2e17c8d87677c6e8cdd043d99b6ee2c"} Feb 27 16:39:25 crc kubenswrapper[4814]: I0227 16:39:25.090374 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r8hf" event={"ID":"c484c9ee-2da0-4ec8-adf2-5ded9c667133","Type":"ContainerStarted","Data":"b0e31a7239e538d1e4782be5c6be5a23afcf41643c722bb9fc248bf507ec165f"} Feb 27 16:39:25 crc kubenswrapper[4814]: I0227 16:39:25.090811 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r8hf" event={"ID":"c484c9ee-2da0-4ec8-adf2-5ded9c667133","Type":"ContainerStarted","Data":"d4e86945b50caf86d451fd70628279b781e4af0504232cba25ce1e55e8d89c15"} Feb 27 16:39:25 crc kubenswrapper[4814]: I0227 16:39:25.090880 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r8hf" event={"ID":"c484c9ee-2da0-4ec8-adf2-5ded9c667133","Type":"ContainerStarted","Data":"1a73f0da72a2e65af3d21ebd9ca5801573946ca36ef8829d965059d82307f9dc"} Feb 27 16:39:25 crc kubenswrapper[4814]: I0227 16:39:25.090892 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r8hf" event={"ID":"c484c9ee-2da0-4ec8-adf2-5ded9c667133","Type":"ContainerStarted","Data":"cb540a0141ed331f1d94a599c7aec2d57596063f1fcf19206368d0cb1a5134a7"} Feb 27 16:39:25 crc kubenswrapper[4814]: I0227 16:39:25.090902 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r8hf" event={"ID":"c484c9ee-2da0-4ec8-adf2-5ded9c667133","Type":"ContainerStarted","Data":"7529a9b44f90522e34444f868492be447c8b2a0d711c214c309f44ce31f15750"} Feb 27 16:39:26 crc kubenswrapper[4814]: I0227 16:39:26.103749 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r8hf" event={"ID":"c484c9ee-2da0-4ec8-adf2-5ded9c667133","Type":"ContainerStarted","Data":"e627da198a9985751fea44e7110169d8784b5a192f4b3d4306dc926ec21a0b88"} Feb 27 16:39:26 crc kubenswrapper[4814]: I0227 16:39:26.105210 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:26 crc kubenswrapper[4814]: I0227 16:39:26.137800 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-5r8hf" podStartSLOduration=6.236060031 podStartE2EDuration="14.137772819s" podCreationTimestamp="2026-02-27 16:39:12 +0000 UTC" firstStartedPulling="2026-02-27 16:39:13.461366309 +0000 UTC m=+965.913991139" lastFinishedPulling="2026-02-27 16:39:21.363079087 +0000 UTC m=+973.815703927" observedRunningTime="2026-02-27 16:39:26.132720203 +0000 UTC m=+978.585345033" watchObservedRunningTime="2026-02-27 16:39:26.137772819 +0000 UTC m=+978.590397689" Feb 27 16:39:28 crc kubenswrapper[4814]: I0227 16:39:28.253281 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:28 crc kubenswrapper[4814]: I0227 16:39:28.323873 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:32 crc kubenswrapper[4814]: I0227 16:39:32.621898 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2bkcm"] Feb 27 16:39:32 crc kubenswrapper[4814]: I0227 16:39:32.632050 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:32 crc kubenswrapper[4814]: I0227 16:39:32.658301 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2bkcm"] Feb 27 16:39:32 crc kubenswrapper[4814]: I0227 16:39:32.709166 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54ab01f3-4f82-42e4-92f6-d13d3c55078e-utilities\") pod \"community-operators-2bkcm\" (UID: \"54ab01f3-4f82-42e4-92f6-d13d3c55078e\") " pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:32 crc kubenswrapper[4814]: I0227 16:39:32.709295 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdszt\" (UniqueName: \"kubernetes.io/projected/54ab01f3-4f82-42e4-92f6-d13d3c55078e-kube-api-access-tdszt\") pod \"community-operators-2bkcm\" (UID: \"54ab01f3-4f82-42e4-92f6-d13d3c55078e\") " pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:32 crc kubenswrapper[4814]: I0227 16:39:32.709337 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54ab01f3-4f82-42e4-92f6-d13d3c55078e-catalog-content\") pod \"community-operators-2bkcm\" (UID: \"54ab01f3-4f82-42e4-92f6-d13d3c55078e\") " pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:32 crc kubenswrapper[4814]: I0227 16:39:32.810513 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54ab01f3-4f82-42e4-92f6-d13d3c55078e-utilities\") pod \"community-operators-2bkcm\" (UID: \"54ab01f3-4f82-42e4-92f6-d13d3c55078e\") " pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:32 crc kubenswrapper[4814]: I0227 16:39:32.810581 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdszt\" (UniqueName: \"kubernetes.io/projected/54ab01f3-4f82-42e4-92f6-d13d3c55078e-kube-api-access-tdszt\") pod \"community-operators-2bkcm\" (UID: \"54ab01f3-4f82-42e4-92f6-d13d3c55078e\") " pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:32 crc kubenswrapper[4814]: I0227 16:39:32.810605 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54ab01f3-4f82-42e4-92f6-d13d3c55078e-catalog-content\") pod \"community-operators-2bkcm\" (UID: \"54ab01f3-4f82-42e4-92f6-d13d3c55078e\") " pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:32 crc kubenswrapper[4814]: I0227 16:39:32.811048 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54ab01f3-4f82-42e4-92f6-d13d3c55078e-catalog-content\") pod \"community-operators-2bkcm\" (UID: \"54ab01f3-4f82-42e4-92f6-d13d3c55078e\") " pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:32 crc kubenswrapper[4814]: I0227 16:39:32.811322 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54ab01f3-4f82-42e4-92f6-d13d3c55078e-utilities\") pod \"community-operators-2bkcm\" (UID: \"54ab01f3-4f82-42e4-92f6-d13d3c55078e\") " pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:32 crc kubenswrapper[4814]: I0227 16:39:32.833339 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdszt\" (UniqueName: \"kubernetes.io/projected/54ab01f3-4f82-42e4-92f6-d13d3c55078e-kube-api-access-tdszt\") pod \"community-operators-2bkcm\" (UID: \"54ab01f3-4f82-42e4-92f6-d13d3c55078e\") " pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:32 crc kubenswrapper[4814]: I0227 16:39:32.972021 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:33 crc kubenswrapper[4814]: I0227 16:39:33.256022 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-lt84m" Feb 27 16:39:33 crc kubenswrapper[4814]: I0227 16:39:33.296895 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2bkcm"] Feb 27 16:39:33 crc kubenswrapper[4814]: I0227 16:39:33.904605 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-86ddb6bd46-5lqjv" Feb 27 16:39:34 crc kubenswrapper[4814]: I0227 16:39:34.172661 4814 generic.go:334] "Generic (PLEG): container finished" podID="54ab01f3-4f82-42e4-92f6-d13d3c55078e" containerID="75a8cacc222b5176b66637b2a9d17bf15c9dff58cf7130648f415d32a1401cef" exitCode=0 Feb 27 16:39:34 crc kubenswrapper[4814]: I0227 16:39:34.172766 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2bkcm" event={"ID":"54ab01f3-4f82-42e4-92f6-d13d3c55078e","Type":"ContainerDied","Data":"75a8cacc222b5176b66637b2a9d17bf15c9dff58cf7130648f415d32a1401cef"} Feb 27 16:39:34 crc kubenswrapper[4814]: I0227 16:39:34.173140 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2bkcm" event={"ID":"54ab01f3-4f82-42e4-92f6-d13d3c55078e","Type":"ContainerStarted","Data":"18b5f4a43aab6abdf27854298459b02f12e93d74f1ab2fd742289865a1a14a7a"} Feb 27 16:39:34 crc kubenswrapper[4814]: I0227 16:39:34.792943 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-v59vj" Feb 27 16:39:35 crc kubenswrapper[4814]: I0227 16:39:35.182298 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2bkcm" event={"ID":"54ab01f3-4f82-42e4-92f6-d13d3c55078e","Type":"ContainerStarted","Data":"2a1d94bfb6c0d10fe14949feaf4325d0afd78d8ff97f148c5ed121f49372f5e2"} Feb 27 16:39:36 crc kubenswrapper[4814]: I0227 16:39:36.192989 4814 generic.go:334] "Generic (PLEG): container finished" podID="54ab01f3-4f82-42e4-92f6-d13d3c55078e" containerID="2a1d94bfb6c0d10fe14949feaf4325d0afd78d8ff97f148c5ed121f49372f5e2" exitCode=0 Feb 27 16:39:36 crc kubenswrapper[4814]: I0227 16:39:36.193038 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2bkcm" event={"ID":"54ab01f3-4f82-42e4-92f6-d13d3c55078e","Type":"ContainerDied","Data":"2a1d94bfb6c0d10fe14949feaf4325d0afd78d8ff97f148c5ed121f49372f5e2"} Feb 27 16:39:37 crc kubenswrapper[4814]: I0227 16:39:37.200732 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2bkcm" event={"ID":"54ab01f3-4f82-42e4-92f6-d13d3c55078e","Type":"ContainerStarted","Data":"3f6fb3479dcb1c703cf1509d1f962972613a3e1bbb938b49ff37997674c98b9e"} Feb 27 16:39:37 crc kubenswrapper[4814]: I0227 16:39:37.217533 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2bkcm" podStartSLOduration=2.803851302 podStartE2EDuration="5.217514944s" podCreationTimestamp="2026-02-27 16:39:32 +0000 UTC" firstStartedPulling="2026-02-27 16:39:34.174284058 +0000 UTC m=+986.626908918" lastFinishedPulling="2026-02-27 16:39:36.58794769 +0000 UTC m=+989.040572560" observedRunningTime="2026-02-27 16:39:37.216106781 +0000 UTC m=+989.668731611" watchObservedRunningTime="2026-02-27 16:39:37.217514944 +0000 UTC m=+989.670139774" Feb 27 16:39:41 crc kubenswrapper[4814]: I0227 16:39:41.192047 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-rnx2c"] Feb 27 16:39:41 crc kubenswrapper[4814]: I0227 16:39:41.193288 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rnx2c" Feb 27 16:39:41 crc kubenswrapper[4814]: I0227 16:39:41.200398 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rnx2c"] Feb 27 16:39:41 crc kubenswrapper[4814]: I0227 16:39:41.201174 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 27 16:39:41 crc kubenswrapper[4814]: I0227 16:39:41.201661 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-g7rhj" Feb 27 16:39:41 crc kubenswrapper[4814]: I0227 16:39:41.201665 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 27 16:39:41 crc kubenswrapper[4814]: I0227 16:39:41.258135 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqphg\" (UniqueName: \"kubernetes.io/projected/f571f89f-5528-4a9c-9ea4-f1449e67eb56-kube-api-access-gqphg\") pod \"openstack-operator-index-rnx2c\" (UID: \"f571f89f-5528-4a9c-9ea4-f1449e67eb56\") " pod="openstack-operators/openstack-operator-index-rnx2c" Feb 27 16:39:41 crc kubenswrapper[4814]: I0227 16:39:41.359714 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqphg\" (UniqueName: \"kubernetes.io/projected/f571f89f-5528-4a9c-9ea4-f1449e67eb56-kube-api-access-gqphg\") pod \"openstack-operator-index-rnx2c\" (UID: \"f571f89f-5528-4a9c-9ea4-f1449e67eb56\") " pod="openstack-operators/openstack-operator-index-rnx2c" Feb 27 16:39:41 crc kubenswrapper[4814]: I0227 16:39:41.380524 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqphg\" (UniqueName: \"kubernetes.io/projected/f571f89f-5528-4a9c-9ea4-f1449e67eb56-kube-api-access-gqphg\") pod \"openstack-operator-index-rnx2c\" (UID: \"f571f89f-5528-4a9c-9ea4-f1449e67eb56\") " pod="openstack-operators/openstack-operator-index-rnx2c" Feb 27 16:39:41 crc kubenswrapper[4814]: I0227 16:39:41.510921 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rnx2c" Feb 27 16:39:42 crc kubenswrapper[4814]: I0227 16:39:42.055985 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rnx2c"] Feb 27 16:39:42 crc kubenswrapper[4814]: I0227 16:39:42.233883 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rnx2c" event={"ID":"f571f89f-5528-4a9c-9ea4-f1449e67eb56","Type":"ContainerStarted","Data":"1a929b8d870e1d0df2177dabf4a61a3ce5d8a3c9fef1767ea0c8c76dbfb81910"} Feb 27 16:39:42 crc kubenswrapper[4814]: I0227 16:39:42.972320 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:42 crc kubenswrapper[4814]: I0227 16:39:42.972673 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:43 crc kubenswrapper[4814]: I0227 16:39:43.025970 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:43 crc kubenswrapper[4814]: I0227 16:39:43.256570 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-5r8hf" Feb 27 16:39:43 crc kubenswrapper[4814]: I0227 16:39:43.303936 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:45 crc kubenswrapper[4814]: I0227 16:39:45.260894 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rnx2c" event={"ID":"f571f89f-5528-4a9c-9ea4-f1449e67eb56","Type":"ContainerStarted","Data":"fb424ecfec03d86062d8c053226f109ad144d72ba0f33e14ad220ec3fd858025"} Feb 27 16:39:45 crc kubenswrapper[4814]: I0227 16:39:45.289246 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-rnx2c" podStartSLOduration=2.119793126 podStartE2EDuration="4.289215564s" podCreationTimestamp="2026-02-27 16:39:41 +0000 UTC" firstStartedPulling="2026-02-27 16:39:42.063957403 +0000 UTC m=+994.516582233" lastFinishedPulling="2026-02-27 16:39:44.233379801 +0000 UTC m=+996.686004671" observedRunningTime="2026-02-27 16:39:45.28584709 +0000 UTC m=+997.738471970" watchObservedRunningTime="2026-02-27 16:39:45.289215564 +0000 UTC m=+997.741840434" Feb 27 16:39:46 crc kubenswrapper[4814]: I0227 16:39:46.986815 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2bkcm"] Feb 27 16:39:46 crc kubenswrapper[4814]: I0227 16:39:46.987199 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2bkcm" podUID="54ab01f3-4f82-42e4-92f6-d13d3c55078e" containerName="registry-server" containerID="cri-o://3f6fb3479dcb1c703cf1509d1f962972613a3e1bbb938b49ff37997674c98b9e" gracePeriod=2 Feb 27 16:39:47 crc kubenswrapper[4814]: I0227 16:39:47.294823 4814 generic.go:334] "Generic (PLEG): container finished" podID="54ab01f3-4f82-42e4-92f6-d13d3c55078e" containerID="3f6fb3479dcb1c703cf1509d1f962972613a3e1bbb938b49ff37997674c98b9e" exitCode=0 Feb 27 16:39:47 crc kubenswrapper[4814]: I0227 16:39:47.295218 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2bkcm" event={"ID":"54ab01f3-4f82-42e4-92f6-d13d3c55078e","Type":"ContainerDied","Data":"3f6fb3479dcb1c703cf1509d1f962972613a3e1bbb938b49ff37997674c98b9e"} Feb 27 16:39:47 crc kubenswrapper[4814]: I0227 16:39:47.418076 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:47 crc kubenswrapper[4814]: I0227 16:39:47.554310 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54ab01f3-4f82-42e4-92f6-d13d3c55078e-catalog-content\") pod \"54ab01f3-4f82-42e4-92f6-d13d3c55078e\" (UID: \"54ab01f3-4f82-42e4-92f6-d13d3c55078e\") " Feb 27 16:39:47 crc kubenswrapper[4814]: I0227 16:39:47.554360 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdszt\" (UniqueName: \"kubernetes.io/projected/54ab01f3-4f82-42e4-92f6-d13d3c55078e-kube-api-access-tdszt\") pod \"54ab01f3-4f82-42e4-92f6-d13d3c55078e\" (UID: \"54ab01f3-4f82-42e4-92f6-d13d3c55078e\") " Feb 27 16:39:47 crc kubenswrapper[4814]: I0227 16:39:47.554425 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54ab01f3-4f82-42e4-92f6-d13d3c55078e-utilities\") pod \"54ab01f3-4f82-42e4-92f6-d13d3c55078e\" (UID: \"54ab01f3-4f82-42e4-92f6-d13d3c55078e\") " Feb 27 16:39:47 crc kubenswrapper[4814]: I0227 16:39:47.555520 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54ab01f3-4f82-42e4-92f6-d13d3c55078e-utilities" (OuterVolumeSpecName: "utilities") pod "54ab01f3-4f82-42e4-92f6-d13d3c55078e" (UID: "54ab01f3-4f82-42e4-92f6-d13d3c55078e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:39:47 crc kubenswrapper[4814]: I0227 16:39:47.561036 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54ab01f3-4f82-42e4-92f6-d13d3c55078e-kube-api-access-tdszt" (OuterVolumeSpecName: "kube-api-access-tdszt") pod "54ab01f3-4f82-42e4-92f6-d13d3c55078e" (UID: "54ab01f3-4f82-42e4-92f6-d13d3c55078e"). InnerVolumeSpecName "kube-api-access-tdszt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:39:47 crc kubenswrapper[4814]: I0227 16:39:47.627038 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54ab01f3-4f82-42e4-92f6-d13d3c55078e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54ab01f3-4f82-42e4-92f6-d13d3c55078e" (UID: "54ab01f3-4f82-42e4-92f6-d13d3c55078e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:39:47 crc kubenswrapper[4814]: I0227 16:39:47.656065 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54ab01f3-4f82-42e4-92f6-d13d3c55078e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:39:47 crc kubenswrapper[4814]: I0227 16:39:47.656102 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdszt\" (UniqueName: \"kubernetes.io/projected/54ab01f3-4f82-42e4-92f6-d13d3c55078e-kube-api-access-tdszt\") on node \"crc\" DevicePath \"\"" Feb 27 16:39:47 crc kubenswrapper[4814]: I0227 16:39:47.656114 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54ab01f3-4f82-42e4-92f6-d13d3c55078e-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:39:48 crc kubenswrapper[4814]: I0227 16:39:48.301634 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2bkcm" event={"ID":"54ab01f3-4f82-42e4-92f6-d13d3c55078e","Type":"ContainerDied","Data":"18b5f4a43aab6abdf27854298459b02f12e93d74f1ab2fd742289865a1a14a7a"} Feb 27 16:39:48 crc kubenswrapper[4814]: I0227 16:39:48.301681 4814 scope.go:117] "RemoveContainer" containerID="3f6fb3479dcb1c703cf1509d1f962972613a3e1bbb938b49ff37997674c98b9e" Feb 27 16:39:48 crc kubenswrapper[4814]: I0227 16:39:48.301777 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2bkcm" Feb 27 16:39:48 crc kubenswrapper[4814]: I0227 16:39:48.324874 4814 scope.go:117] "RemoveContainer" containerID="2a1d94bfb6c0d10fe14949feaf4325d0afd78d8ff97f148c5ed121f49372f5e2" Feb 27 16:39:48 crc kubenswrapper[4814]: I0227 16:39:48.339167 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2bkcm"] Feb 27 16:39:48 crc kubenswrapper[4814]: I0227 16:39:48.342647 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2bkcm"] Feb 27 16:39:48 crc kubenswrapper[4814]: I0227 16:39:48.363789 4814 scope.go:117] "RemoveContainer" containerID="75a8cacc222b5176b66637b2a9d17bf15c9dff58cf7130648f415d32a1401cef" Feb 27 16:39:48 crc kubenswrapper[4814]: I0227 16:39:48.495016 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54ab01f3-4f82-42e4-92f6-d13d3c55078e" path="/var/lib/kubelet/pods/54ab01f3-4f82-42e4-92f6-d13d3c55078e/volumes" Feb 27 16:39:51 crc kubenswrapper[4814]: I0227 16:39:51.511999 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-rnx2c" Feb 27 16:39:51 crc kubenswrapper[4814]: I0227 16:39:51.512080 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-rnx2c" Feb 27 16:39:51 crc kubenswrapper[4814]: I0227 16:39:51.558326 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-rnx2c" Feb 27 16:39:52 crc kubenswrapper[4814]: I0227 16:39:52.364230 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-rnx2c" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.648557 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g"] Feb 27 16:39:53 crc kubenswrapper[4814]: E0227 16:39:53.649023 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54ab01f3-4f82-42e4-92f6-d13d3c55078e" containerName="extract-utilities" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.649055 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="54ab01f3-4f82-42e4-92f6-d13d3c55078e" containerName="extract-utilities" Feb 27 16:39:53 crc kubenswrapper[4814]: E0227 16:39:53.649089 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54ab01f3-4f82-42e4-92f6-d13d3c55078e" containerName="registry-server" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.649106 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="54ab01f3-4f82-42e4-92f6-d13d3c55078e" containerName="registry-server" Feb 27 16:39:53 crc kubenswrapper[4814]: E0227 16:39:53.649167 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54ab01f3-4f82-42e4-92f6-d13d3c55078e" containerName="extract-content" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.649187 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="54ab01f3-4f82-42e4-92f6-d13d3c55078e" containerName="extract-content" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.649524 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="54ab01f3-4f82-42e4-92f6-d13d3c55078e" containerName="registry-server" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.651502 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.655306 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g"] Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.657147 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-x6qnk" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.733496 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwwks\" (UniqueName: \"kubernetes.io/projected/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-kube-api-access-wwwks\") pod \"c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g\" (UID: \"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8\") " pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.733871 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-util\") pod \"c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g\" (UID: \"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8\") " pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.733900 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-bundle\") pod \"c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g\" (UID: \"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8\") " pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.835718 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwwks\" (UniqueName: \"kubernetes.io/projected/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-kube-api-access-wwwks\") pod \"c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g\" (UID: \"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8\") " pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.835938 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-util\") pod \"c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g\" (UID: \"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8\") " pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.835997 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-bundle\") pod \"c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g\" (UID: \"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8\") " pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.836722 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-bundle\") pod \"c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g\" (UID: \"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8\") " pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.836821 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-util\") pod \"c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g\" (UID: \"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8\") " pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.859373 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwwks\" (UniqueName: \"kubernetes.io/projected/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-kube-api-access-wwwks\") pod \"c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g\" (UID: \"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8\") " pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" Feb 27 16:39:53 crc kubenswrapper[4814]: I0227 16:39:53.982894 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" Feb 27 16:39:54 crc kubenswrapper[4814]: I0227 16:39:54.499556 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g"] Feb 27 16:39:55 crc kubenswrapper[4814]: I0227 16:39:55.361841 4814 generic.go:334] "Generic (PLEG): container finished" podID="0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8" containerID="40b85fd2dcc4b1d73f2e94015fea9d8593f837adb8b659a78052b0dbd7aaa21b" exitCode=0 Feb 27 16:39:55 crc kubenswrapper[4814]: I0227 16:39:55.361914 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" event={"ID":"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8","Type":"ContainerDied","Data":"40b85fd2dcc4b1d73f2e94015fea9d8593f837adb8b659a78052b0dbd7aaa21b"} Feb 27 16:39:55 crc kubenswrapper[4814]: I0227 16:39:55.362340 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" event={"ID":"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8","Type":"ContainerStarted","Data":"dd210061619e8f1aed4e42e017f9ccdc41d5efe2be538dcc8be3ce91ac6d2f23"} Feb 27 16:39:58 crc kubenswrapper[4814]: I0227 16:39:58.387518 4814 generic.go:334] "Generic (PLEG): container finished" podID="0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8" containerID="d03a343ba2bc227a8515e9685196df88407ca8936a92b7db2d71b6ef43868854" exitCode=0 Feb 27 16:39:58 crc kubenswrapper[4814]: I0227 16:39:58.387594 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" event={"ID":"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8","Type":"ContainerDied","Data":"d03a343ba2bc227a8515e9685196df88407ca8936a92b7db2d71b6ef43868854"} Feb 27 16:39:59 crc kubenswrapper[4814]: I0227 16:39:59.399411 4814 generic.go:334] "Generic (PLEG): container finished" podID="0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8" containerID="8776b397f83d7b5884dec04023a9c87ac19e933ac6b7f1ed5605ed218c2a237a" exitCode=0 Feb 27 16:39:59 crc kubenswrapper[4814]: I0227 16:39:59.399481 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" event={"ID":"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8","Type":"ContainerDied","Data":"8776b397f83d7b5884dec04023a9c87ac19e933ac6b7f1ed5605ed218c2a237a"} Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.139846 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536840-rgdnq"] Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.141478 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536840-rgdnq" Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.144668 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.144996 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.145439 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.153957 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536840-rgdnq"] Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.236510 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqwr8\" (UniqueName: \"kubernetes.io/projected/280aeb00-70d0-4981-a5ee-430bf4014055-kube-api-access-bqwr8\") pod \"auto-csr-approver-29536840-rgdnq\" (UID: \"280aeb00-70d0-4981-a5ee-430bf4014055\") " pod="openshift-infra/auto-csr-approver-29536840-rgdnq" Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.338124 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqwr8\" (UniqueName: \"kubernetes.io/projected/280aeb00-70d0-4981-a5ee-430bf4014055-kube-api-access-bqwr8\") pod \"auto-csr-approver-29536840-rgdnq\" (UID: \"280aeb00-70d0-4981-a5ee-430bf4014055\") " pod="openshift-infra/auto-csr-approver-29536840-rgdnq" Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.360557 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqwr8\" (UniqueName: \"kubernetes.io/projected/280aeb00-70d0-4981-a5ee-430bf4014055-kube-api-access-bqwr8\") pod \"auto-csr-approver-29536840-rgdnq\" (UID: \"280aeb00-70d0-4981-a5ee-430bf4014055\") " pod="openshift-infra/auto-csr-approver-29536840-rgdnq" Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.471682 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536840-rgdnq" Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.712517 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536840-rgdnq"] Feb 27 16:40:00 crc kubenswrapper[4814]: W0227 16:40:00.724190 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod280aeb00_70d0_4981_a5ee_430bf4014055.slice/crio-17367e805a4f568abf94cfb7130540eb9b733449892922ba9fea5d4f70d932dc WatchSource:0}: Error finding container 17367e805a4f568abf94cfb7130540eb9b733449892922ba9fea5d4f70d932dc: Status 404 returned error can't find the container with id 17367e805a4f568abf94cfb7130540eb9b733449892922ba9fea5d4f70d932dc Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.770606 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.846812 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-util\") pod \"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8\" (UID: \"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8\") " Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.846889 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwwks\" (UniqueName: \"kubernetes.io/projected/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-kube-api-access-wwwks\") pod \"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8\" (UID: \"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8\") " Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.846970 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-bundle\") pod \"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8\" (UID: \"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8\") " Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.848246 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-bundle" (OuterVolumeSpecName: "bundle") pod "0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8" (UID: "0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.853180 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-kube-api-access-wwwks" (OuterVolumeSpecName: "kube-api-access-wwwks") pod "0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8" (UID: "0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8"). InnerVolumeSpecName "kube-api-access-wwwks". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.858513 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-util" (OuterVolumeSpecName: "util") pod "0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8" (UID: "0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.948786 4814 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.948820 4814 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-util\") on node \"crc\" DevicePath \"\"" Feb 27 16:40:00 crc kubenswrapper[4814]: I0227 16:40:00.948829 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwwks\" (UniqueName: \"kubernetes.io/projected/0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8-kube-api-access-wwwks\") on node \"crc\" DevicePath \"\"" Feb 27 16:40:01 crc kubenswrapper[4814]: I0227 16:40:01.412195 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" event={"ID":"0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8","Type":"ContainerDied","Data":"dd210061619e8f1aed4e42e017f9ccdc41d5efe2be538dcc8be3ce91ac6d2f23"} Feb 27 16:40:01 crc kubenswrapper[4814]: I0227 16:40:01.412548 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd210061619e8f1aed4e42e017f9ccdc41d5efe2be538dcc8be3ce91ac6d2f23" Feb 27 16:40:01 crc kubenswrapper[4814]: I0227 16:40:01.412401 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g" Feb 27 16:40:01 crc kubenswrapper[4814]: I0227 16:40:01.413346 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536840-rgdnq" event={"ID":"280aeb00-70d0-4981-a5ee-430bf4014055","Type":"ContainerStarted","Data":"17367e805a4f568abf94cfb7130540eb9b733449892922ba9fea5d4f70d932dc"} Feb 27 16:40:02 crc kubenswrapper[4814]: I0227 16:40:02.427422 4814 generic.go:334] "Generic (PLEG): container finished" podID="280aeb00-70d0-4981-a5ee-430bf4014055" containerID="3bba3d181a7796362e15415c9ce922f887f8662bee51b24906b208723a196e50" exitCode=0 Feb 27 16:40:02 crc kubenswrapper[4814]: I0227 16:40:02.427519 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536840-rgdnq" event={"ID":"280aeb00-70d0-4981-a5ee-430bf4014055","Type":"ContainerDied","Data":"3bba3d181a7796362e15415c9ce922f887f8662bee51b24906b208723a196e50"} Feb 27 16:40:03 crc kubenswrapper[4814]: I0227 16:40:03.808320 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536840-rgdnq" Feb 27 16:40:03 crc kubenswrapper[4814]: I0227 16:40:03.892848 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqwr8\" (UniqueName: \"kubernetes.io/projected/280aeb00-70d0-4981-a5ee-430bf4014055-kube-api-access-bqwr8\") pod \"280aeb00-70d0-4981-a5ee-430bf4014055\" (UID: \"280aeb00-70d0-4981-a5ee-430bf4014055\") " Feb 27 16:40:03 crc kubenswrapper[4814]: I0227 16:40:03.898819 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/280aeb00-70d0-4981-a5ee-430bf4014055-kube-api-access-bqwr8" (OuterVolumeSpecName: "kube-api-access-bqwr8") pod "280aeb00-70d0-4981-a5ee-430bf4014055" (UID: "280aeb00-70d0-4981-a5ee-430bf4014055"). InnerVolumeSpecName "kube-api-access-bqwr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:40:03 crc kubenswrapper[4814]: I0227 16:40:03.994146 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqwr8\" (UniqueName: \"kubernetes.io/projected/280aeb00-70d0-4981-a5ee-430bf4014055-kube-api-access-bqwr8\") on node \"crc\" DevicePath \"\"" Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.444768 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536840-rgdnq" event={"ID":"280aeb00-70d0-4981-a5ee-430bf4014055","Type":"ContainerDied","Data":"17367e805a4f568abf94cfb7130540eb9b733449892922ba9fea5d4f70d932dc"} Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.444805 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17367e805a4f568abf94cfb7130540eb9b733449892922ba9fea5d4f70d932dc" Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.444854 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536840-rgdnq" Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.759891 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-7d5f5556b6-b9hqd"] Feb 27 16:40:04 crc kubenswrapper[4814]: E0227 16:40:04.760291 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="280aeb00-70d0-4981-a5ee-430bf4014055" containerName="oc" Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.760313 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="280aeb00-70d0-4981-a5ee-430bf4014055" containerName="oc" Feb 27 16:40:04 crc kubenswrapper[4814]: E0227 16:40:04.760339 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8" containerName="extract" Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.760353 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8" containerName="extract" Feb 27 16:40:04 crc kubenswrapper[4814]: E0227 16:40:04.760392 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8" containerName="util" Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.760407 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8" containerName="util" Feb 27 16:40:04 crc kubenswrapper[4814]: E0227 16:40:04.760430 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8" containerName="pull" Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.760443 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8" containerName="pull" Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.760661 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="280aeb00-70d0-4981-a5ee-430bf4014055" containerName="oc" Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.760701 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8" containerName="extract" Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.761401 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-7d5f5556b6-b9hqd" Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.763620 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-r9xtw" Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.783035 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-7d5f5556b6-b9hqd"] Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.805055 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bdwt\" (UniqueName: \"kubernetes.io/projected/43af9a66-30c9-4410-862d-e07c38625ea7-kube-api-access-8bdwt\") pod \"openstack-operator-controller-init-7d5f5556b6-b9hqd\" (UID: \"43af9a66-30c9-4410-862d-e07c38625ea7\") " pod="openstack-operators/openstack-operator-controller-init-7d5f5556b6-b9hqd" Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.890625 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536834-w2h2s"] Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.896385 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536834-w2h2s"] Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.905950 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bdwt\" (UniqueName: \"kubernetes.io/projected/43af9a66-30c9-4410-862d-e07c38625ea7-kube-api-access-8bdwt\") pod \"openstack-operator-controller-init-7d5f5556b6-b9hqd\" (UID: \"43af9a66-30c9-4410-862d-e07c38625ea7\") " pod="openstack-operators/openstack-operator-controller-init-7d5f5556b6-b9hqd" Feb 27 16:40:04 crc kubenswrapper[4814]: I0227 16:40:04.924805 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bdwt\" (UniqueName: \"kubernetes.io/projected/43af9a66-30c9-4410-862d-e07c38625ea7-kube-api-access-8bdwt\") pod \"openstack-operator-controller-init-7d5f5556b6-b9hqd\" (UID: \"43af9a66-30c9-4410-862d-e07c38625ea7\") " pod="openstack-operators/openstack-operator-controller-init-7d5f5556b6-b9hqd" Feb 27 16:40:05 crc kubenswrapper[4814]: I0227 16:40:05.079165 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-7d5f5556b6-b9hqd" Feb 27 16:40:05 crc kubenswrapper[4814]: I0227 16:40:05.524867 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-7d5f5556b6-b9hqd"] Feb 27 16:40:05 crc kubenswrapper[4814]: W0227 16:40:05.531389 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43af9a66_30c9_4410_862d_e07c38625ea7.slice/crio-b4539b73d12718bba3dc0c94e944b79a2df391289c118c7dbd7833b4bf6c0fee WatchSource:0}: Error finding container b4539b73d12718bba3dc0c94e944b79a2df391289c118c7dbd7833b4bf6c0fee: Status 404 returned error can't find the container with id b4539b73d12718bba3dc0c94e944b79a2df391289c118c7dbd7833b4bf6c0fee Feb 27 16:40:06 crc kubenswrapper[4814]: I0227 16:40:06.461517 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-7d5f5556b6-b9hqd" event={"ID":"43af9a66-30c9-4410-862d-e07c38625ea7","Type":"ContainerStarted","Data":"b4539b73d12718bba3dc0c94e944b79a2df391289c118c7dbd7833b4bf6c0fee"} Feb 27 16:40:06 crc kubenswrapper[4814]: I0227 16:40:06.496628 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7ca68cb-43ee-405b-a9bf-8a6dc109ee27" path="/var/lib/kubelet/pods/d7ca68cb-43ee-405b-a9bf-8a6dc109ee27/volumes" Feb 27 16:40:09 crc kubenswrapper[4814]: I0227 16:40:09.489635 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-7d5f5556b6-b9hqd" event={"ID":"43af9a66-30c9-4410-862d-e07c38625ea7","Type":"ContainerStarted","Data":"b05c6be1ae990c6303a8876b24860391319c7dd3b0cc5d591ea0b506ee7ff319"} Feb 27 16:40:09 crc kubenswrapper[4814]: I0227 16:40:09.490107 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-7d5f5556b6-b9hqd" Feb 27 16:40:09 crc kubenswrapper[4814]: I0227 16:40:09.546602 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-7d5f5556b6-b9hqd" podStartSLOduration=1.984261859 podStartE2EDuration="5.546579714s" podCreationTimestamp="2026-02-27 16:40:04 +0000 UTC" firstStartedPulling="2026-02-27 16:40:05.533090987 +0000 UTC m=+1017.985715827" lastFinishedPulling="2026-02-27 16:40:09.095408842 +0000 UTC m=+1021.548033682" observedRunningTime="2026-02-27 16:40:09.538760271 +0000 UTC m=+1021.991385131" watchObservedRunningTime="2026-02-27 16:40:09.546579714 +0000 UTC m=+1021.999204554" Feb 27 16:40:15 crc kubenswrapper[4814]: I0227 16:40:15.085477 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-7d5f5556b6-b9hqd" Feb 27 16:40:23 crc kubenswrapper[4814]: I0227 16:40:23.699934 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nvzj5"] Feb 27 16:40:23 crc kubenswrapper[4814]: I0227 16:40:23.701438 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:23 crc kubenswrapper[4814]: I0227 16:40:23.711800 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nvzj5"] Feb 27 16:40:23 crc kubenswrapper[4814]: I0227 16:40:23.795130 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afffcacd-d821-4cfd-aad8-462aafd2986f-catalog-content\") pod \"certified-operators-nvzj5\" (UID: \"afffcacd-d821-4cfd-aad8-462aafd2986f\") " pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:23 crc kubenswrapper[4814]: I0227 16:40:23.795193 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afffcacd-d821-4cfd-aad8-462aafd2986f-utilities\") pod \"certified-operators-nvzj5\" (UID: \"afffcacd-d821-4cfd-aad8-462aafd2986f\") " pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:23 crc kubenswrapper[4814]: I0227 16:40:23.795303 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5rjt\" (UniqueName: \"kubernetes.io/projected/afffcacd-d821-4cfd-aad8-462aafd2986f-kube-api-access-s5rjt\") pod \"certified-operators-nvzj5\" (UID: \"afffcacd-d821-4cfd-aad8-462aafd2986f\") " pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:23 crc kubenswrapper[4814]: I0227 16:40:23.896769 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5rjt\" (UniqueName: \"kubernetes.io/projected/afffcacd-d821-4cfd-aad8-462aafd2986f-kube-api-access-s5rjt\") pod \"certified-operators-nvzj5\" (UID: \"afffcacd-d821-4cfd-aad8-462aafd2986f\") " pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:23 crc kubenswrapper[4814]: I0227 16:40:23.896843 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afffcacd-d821-4cfd-aad8-462aafd2986f-catalog-content\") pod \"certified-operators-nvzj5\" (UID: \"afffcacd-d821-4cfd-aad8-462aafd2986f\") " pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:23 crc kubenswrapper[4814]: I0227 16:40:23.896871 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afffcacd-d821-4cfd-aad8-462aafd2986f-utilities\") pod \"certified-operators-nvzj5\" (UID: \"afffcacd-d821-4cfd-aad8-462aafd2986f\") " pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:23 crc kubenswrapper[4814]: I0227 16:40:23.897867 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afffcacd-d821-4cfd-aad8-462aafd2986f-utilities\") pod \"certified-operators-nvzj5\" (UID: \"afffcacd-d821-4cfd-aad8-462aafd2986f\") " pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:23 crc kubenswrapper[4814]: I0227 16:40:23.897906 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afffcacd-d821-4cfd-aad8-462aafd2986f-catalog-content\") pod \"certified-operators-nvzj5\" (UID: \"afffcacd-d821-4cfd-aad8-462aafd2986f\") " pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:23 crc kubenswrapper[4814]: I0227 16:40:23.918154 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5rjt\" (UniqueName: \"kubernetes.io/projected/afffcacd-d821-4cfd-aad8-462aafd2986f-kube-api-access-s5rjt\") pod \"certified-operators-nvzj5\" (UID: \"afffcacd-d821-4cfd-aad8-462aafd2986f\") " pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:24 crc kubenswrapper[4814]: I0227 16:40:24.020959 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:24 crc kubenswrapper[4814]: I0227 16:40:24.344849 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nvzj5"] Feb 27 16:40:24 crc kubenswrapper[4814]: W0227 16:40:24.351525 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafffcacd_d821_4cfd_aad8_462aafd2986f.slice/crio-de4ed563d472ca6c817f216145bff88ec6d1c30ea8c10eb398e463dbebd74197 WatchSource:0}: Error finding container de4ed563d472ca6c817f216145bff88ec6d1c30ea8c10eb398e463dbebd74197: Status 404 returned error can't find the container with id de4ed563d472ca6c817f216145bff88ec6d1c30ea8c10eb398e463dbebd74197 Feb 27 16:40:24 crc kubenswrapper[4814]: I0227 16:40:24.596533 4814 generic.go:334] "Generic (PLEG): container finished" podID="afffcacd-d821-4cfd-aad8-462aafd2986f" containerID="2aee2051d874c3163014ff09ec4d518d320944c0447ff89117be8323300176e5" exitCode=0 Feb 27 16:40:24 crc kubenswrapper[4814]: I0227 16:40:24.596621 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nvzj5" event={"ID":"afffcacd-d821-4cfd-aad8-462aafd2986f","Type":"ContainerDied","Data":"2aee2051d874c3163014ff09ec4d518d320944c0447ff89117be8323300176e5"} Feb 27 16:40:24 crc kubenswrapper[4814]: I0227 16:40:24.597217 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nvzj5" event={"ID":"afffcacd-d821-4cfd-aad8-462aafd2986f","Type":"ContainerStarted","Data":"de4ed563d472ca6c817f216145bff88ec6d1c30ea8c10eb398e463dbebd74197"} Feb 27 16:40:25 crc kubenswrapper[4814]: I0227 16:40:25.603471 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nvzj5" event={"ID":"afffcacd-d821-4cfd-aad8-462aafd2986f","Type":"ContainerStarted","Data":"9d3f7a045fdbb41ede0f5050fc4f55a92ead8ca0c72ec9718d0151a396028245"} Feb 27 16:40:26 crc kubenswrapper[4814]: I0227 16:40:26.376830 4814 scope.go:117] "RemoveContainer" containerID="937fa75707ae28a50ee5d55cfc7a9b9dfa99e3bc62026ff27ce4d0a8f6c024d6" Feb 27 16:40:26 crc kubenswrapper[4814]: I0227 16:40:26.612276 4814 generic.go:334] "Generic (PLEG): container finished" podID="afffcacd-d821-4cfd-aad8-462aafd2986f" containerID="9d3f7a045fdbb41ede0f5050fc4f55a92ead8ca0c72ec9718d0151a396028245" exitCode=0 Feb 27 16:40:26 crc kubenswrapper[4814]: I0227 16:40:26.612441 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nvzj5" event={"ID":"afffcacd-d821-4cfd-aad8-462aafd2986f","Type":"ContainerDied","Data":"9d3f7a045fdbb41ede0f5050fc4f55a92ead8ca0c72ec9718d0151a396028245"} Feb 27 16:40:27 crc kubenswrapper[4814]: I0227 16:40:27.620624 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nvzj5" event={"ID":"afffcacd-d821-4cfd-aad8-462aafd2986f","Type":"ContainerStarted","Data":"73a1b5d0586039162b6eb935fdfed5ecf01f42d81a7bd43b2f1bf99c52420a28"} Feb 27 16:40:27 crc kubenswrapper[4814]: I0227 16:40:27.638988 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nvzj5" podStartSLOduration=2.044081375 podStartE2EDuration="4.63897071s" podCreationTimestamp="2026-02-27 16:40:23 +0000 UTC" firstStartedPulling="2026-02-27 16:40:24.598245528 +0000 UTC m=+1037.050870348" lastFinishedPulling="2026-02-27 16:40:27.193134863 +0000 UTC m=+1039.645759683" observedRunningTime="2026-02-27 16:40:27.637466545 +0000 UTC m=+1040.090091375" watchObservedRunningTime="2026-02-27 16:40:27.63897071 +0000 UTC m=+1040.091595540" Feb 27 16:40:34 crc kubenswrapper[4814]: I0227 16:40:34.021155 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:34 crc kubenswrapper[4814]: I0227 16:40:34.021738 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:34 crc kubenswrapper[4814]: I0227 16:40:34.082993 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:34 crc kubenswrapper[4814]: I0227 16:40:34.753768 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:34 crc kubenswrapper[4814]: I0227 16:40:34.806469 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nvzj5"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.714497 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nvzj5" podUID="afffcacd-d821-4cfd-aad8-462aafd2986f" containerName="registry-server" containerID="cri-o://73a1b5d0586039162b6eb935fdfed5ecf01f42d81a7bd43b2f1bf99c52420a28" gracePeriod=2 Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.719017 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6db6876945-xnhhg"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.720189 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-xnhhg" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.724872 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-29zgc" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.735031 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-2d2n7"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.736007 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-2d2n7" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.737397 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-hnsmv" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.759348 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6db6876945-xnhhg"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.766386 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-2d2n7"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.779515 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-5d87c9d997-pjncz"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.780539 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-pjncz" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.784466 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-64db6967f8-9kqgh"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.785219 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-9kqgh" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.785528 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-vt72q" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.788886 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-5d87c9d997-pjncz"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.789203 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-7z48b" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.796000 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b8zt\" (UniqueName: \"kubernetes.io/projected/de6e87db-4283-4aab-a97e-23792fd3b480-kube-api-access-2b8zt\") pod \"cinder-operator-controller-manager-55d77d7b5c-2d2n7\" (UID: \"de6e87db-4283-4aab-a97e-23792fd3b480\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-2d2n7" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.796227 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d8zb\" (UniqueName: \"kubernetes.io/projected/6ede514a-7a77-4cfe-a746-3d7bd2f5d7fd-kube-api-access-6d8zb\") pod \"barbican-operator-controller-manager-6db6876945-xnhhg\" (UID: \"6ede514a-7a77-4cfe-a746-3d7bd2f5d7fd\") " pod="openstack-operators/barbican-operator-controller-manager-6db6876945-xnhhg" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.806896 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-64db6967f8-9kqgh"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.817207 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-cf99c678f-ccrlj"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.818585 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-ccrlj" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.821093 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-s8fl7" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.830592 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-cf99c678f-ccrlj"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.838328 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-p74f5"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.841455 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-p74f5" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.855545 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-4pbgc" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.883402 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-p74f5"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.897823 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p7xw\" (UniqueName: \"kubernetes.io/projected/6ee0f0c6-52ee-4d20-999f-a0a579dcdaef-kube-api-access-2p7xw\") pod \"designate-operator-controller-manager-5d87c9d997-pjncz\" (UID: \"6ee0f0c6-52ee-4d20-999f-a0a579dcdaef\") " pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-pjncz" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.898108 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcwsc\" (UniqueName: \"kubernetes.io/projected/a3038d2f-e4de-4d0f-8ac0-e5c666077395-kube-api-access-pcwsc\") pod \"heat-operator-controller-manager-cf99c678f-ccrlj\" (UID: \"a3038d2f-e4de-4d0f-8ac0-e5c666077395\") " pod="openstack-operators/heat-operator-controller-manager-cf99c678f-ccrlj" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.898187 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d8zb\" (UniqueName: \"kubernetes.io/projected/6ede514a-7a77-4cfe-a746-3d7bd2f5d7fd-kube-api-access-6d8zb\") pod \"barbican-operator-controller-manager-6db6876945-xnhhg\" (UID: \"6ede514a-7a77-4cfe-a746-3d7bd2f5d7fd\") " pod="openstack-operators/barbican-operator-controller-manager-6db6876945-xnhhg" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.898347 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46gnd\" (UniqueName: \"kubernetes.io/projected/ea6ecef0-2796-4c32-9a9a-ee178574d2f9-kube-api-access-46gnd\") pod \"horizon-operator-controller-manager-78bc7f9bd9-p74f5\" (UID: \"ea6ecef0-2796-4c32-9a9a-ee178574d2f9\") " pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-p74f5" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.898444 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b8zt\" (UniqueName: \"kubernetes.io/projected/de6e87db-4283-4aab-a97e-23792fd3b480-kube-api-access-2b8zt\") pod \"cinder-operator-controller-manager-55d77d7b5c-2d2n7\" (UID: \"de6e87db-4283-4aab-a97e-23792fd3b480\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-2d2n7" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.898522 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgx42\" (UniqueName: \"kubernetes.io/projected/49b03138-9b4c-486c-af91-37c16e8c2536-kube-api-access-qgx42\") pod \"glance-operator-controller-manager-64db6967f8-9kqgh\" (UID: \"49b03138-9b4c-486c-af91-37c16e8c2536\") " pod="openstack-operators/glance-operator-controller-manager-64db6967f8-9kqgh" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.900489 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.901194 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.903037 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.903078 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-w4vqz" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.920058 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.924044 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-545456dc4-fp474"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.925072 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-fp474" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.929065 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-hgcdh" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.932286 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55ffd4876b-nxrtn"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.932973 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-nxrtn" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.935569 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-67z7d" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.939303 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d8zb\" (UniqueName: \"kubernetes.io/projected/6ede514a-7a77-4cfe-a746-3d7bd2f5d7fd-kube-api-access-6d8zb\") pod \"barbican-operator-controller-manager-6db6876945-xnhhg\" (UID: \"6ede514a-7a77-4cfe-a746-3d7bd2f5d7fd\") " pod="openstack-operators/barbican-operator-controller-manager-6db6876945-xnhhg" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.943396 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-545456dc4-fp474"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.960245 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-c5mcj"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.960999 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-67d996989d-c5mcj" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.961766 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b8zt\" (UniqueName: \"kubernetes.io/projected/de6e87db-4283-4aab-a97e-23792fd3b480-kube-api-access-2b8zt\") pod \"cinder-operator-controller-manager-55d77d7b5c-2d2n7\" (UID: \"de6e87db-4283-4aab-a97e-23792fd3b480\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-2d2n7" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.962459 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-ln268" Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.985908 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55ffd4876b-nxrtn"] Feb 27 16:40:36 crc kubenswrapper[4814]: I0227 16:40:36.991665 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-c5mcj"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.001824 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-556b8b874-f4dv9"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.003111 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-f4dv9" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.003355 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p7xw\" (UniqueName: \"kubernetes.io/projected/6ee0f0c6-52ee-4d20-999f-a0a579dcdaef-kube-api-access-2p7xw\") pod \"designate-operator-controller-manager-5d87c9d997-pjncz\" (UID: \"6ee0f0c6-52ee-4d20-999f-a0a579dcdaef\") " pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-pjncz" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.003420 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcwsc\" (UniqueName: \"kubernetes.io/projected/a3038d2f-e4de-4d0f-8ac0-e5c666077395-kube-api-access-pcwsc\") pod \"heat-operator-controller-manager-cf99c678f-ccrlj\" (UID: \"a3038d2f-e4de-4d0f-8ac0-e5c666077395\") " pod="openstack-operators/heat-operator-controller-manager-cf99c678f-ccrlj" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.003462 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-qc48n\" (UID: \"2d989d20-0a79-448b-8b73-ad9378119232\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.003491 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46gnd\" (UniqueName: \"kubernetes.io/projected/ea6ecef0-2796-4c32-9a9a-ee178574d2f9-kube-api-access-46gnd\") pod \"horizon-operator-controller-manager-78bc7f9bd9-p74f5\" (UID: \"ea6ecef0-2796-4c32-9a9a-ee178574d2f9\") " pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-p74f5" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.003519 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgx42\" (UniqueName: \"kubernetes.io/projected/49b03138-9b4c-486c-af91-37c16e8c2536-kube-api-access-qgx42\") pod \"glance-operator-controller-manager-64db6967f8-9kqgh\" (UID: \"49b03138-9b4c-486c-af91-37c16e8c2536\") " pod="openstack-operators/glance-operator-controller-manager-64db6967f8-9kqgh" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.003537 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkbwh\" (UniqueName: \"kubernetes.io/projected/2192fb73-bd5d-440f-8c01-2871658164d9-kube-api-access-rkbwh\") pod \"ironic-operator-controller-manager-545456dc4-fp474\" (UID: \"2192fb73-bd5d-440f-8c01-2871658164d9\") " pod="openstack-operators/ironic-operator-controller-manager-545456dc4-fp474" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.003554 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz9lb\" (UniqueName: \"kubernetes.io/projected/9b633770-3463-40ee-b5fc-48ec5fb4846c-kube-api-access-wz9lb\") pod \"keystone-operator-controller-manager-55ffd4876b-nxrtn\" (UID: \"9b633770-3463-40ee-b5fc-48ec5fb4846c\") " pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-nxrtn" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.003573 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5n9s\" (UniqueName: \"kubernetes.io/projected/2d989d20-0a79-448b-8b73-ad9378119232-kube-api-access-t5n9s\") pod \"infra-operator-controller-manager-f7fcc58b9-qc48n\" (UID: \"2d989d20-0a79-448b-8b73-ad9378119232\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.007397 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-pw98l" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.009372 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-556b8b874-f4dv9"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.025490 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54688575f-q2gkq"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.026562 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-54688575f-q2gkq" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.031778 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-74b6b5dc96-m5zqv"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.032642 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-9ftr6" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.033217 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54688575f-q2gkq"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.033545 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-m5zqv" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.034580 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p7xw\" (UniqueName: \"kubernetes.io/projected/6ee0f0c6-52ee-4d20-999f-a0a579dcdaef-kube-api-access-2p7xw\") pod \"designate-operator-controller-manager-5d87c9d997-pjncz\" (UID: \"6ee0f0c6-52ee-4d20-999f-a0a579dcdaef\") " pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-pjncz" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.035633 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-xrh2k" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.039496 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcwsc\" (UniqueName: \"kubernetes.io/projected/a3038d2f-e4de-4d0f-8ac0-e5c666077395-kube-api-access-pcwsc\") pod \"heat-operator-controller-manager-cf99c678f-ccrlj\" (UID: \"a3038d2f-e4de-4d0f-8ac0-e5c666077395\") " pod="openstack-operators/heat-operator-controller-manager-cf99c678f-ccrlj" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.040997 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46gnd\" (UniqueName: \"kubernetes.io/projected/ea6ecef0-2796-4c32-9a9a-ee178574d2f9-kube-api-access-46gnd\") pod \"horizon-operator-controller-manager-78bc7f9bd9-p74f5\" (UID: \"ea6ecef0-2796-4c32-9a9a-ee178574d2f9\") " pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-p74f5" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.047317 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-4m2wv"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.048127 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-4m2wv" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.048419 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-74b6b5dc96-m5zqv"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.050525 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgx42\" (UniqueName: \"kubernetes.io/projected/49b03138-9b4c-486c-af91-37c16e8c2536-kube-api-access-qgx42\") pod \"glance-operator-controller-manager-64db6967f8-9kqgh\" (UID: \"49b03138-9b4c-486c-af91-37c16e8c2536\") " pod="openstack-operators/glance-operator-controller-manager-64db6967f8-9kqgh" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.052566 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-qlws2" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.052953 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-xnhhg" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.056791 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-4m2wv"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.070501 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.071365 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.073678 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.074006 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-82tsb" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.074707 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-2d2n7" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.080140 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-75684d597f-5958h"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.081228 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-5958h" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.085058 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-75684d597f-5958h"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.090846 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.096977 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-nxjhr" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.099418 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-648564c9fc-hbdxk"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.100439 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-hbdxk" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.102622 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-pjncz" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.105060 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8cdq\" (UniqueName: \"kubernetes.io/projected/dfcfffcb-36da-4853-8ed7-ba76caba9d3c-kube-api-access-s8cdq\") pod \"manila-operator-controller-manager-67d996989d-c5mcj\" (UID: \"dfcfffcb-36da-4853-8ed7-ba76caba9d3c\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-c5mcj" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.105124 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-qc48n\" (UID: \"2d989d20-0a79-448b-8b73-ad9378119232\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.105152 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qszk7\" (UniqueName: \"kubernetes.io/projected/75ee2f25-f59a-4a60-8245-1b638957d234-kube-api-access-qszk7\") pod \"octavia-operator-controller-manager-5d86c7ddb7-4m2wv\" (UID: \"75ee2f25-f59a-4a60-8245-1b638957d234\") " pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-4m2wv" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.105480 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8hm5\" (UniqueName: \"kubernetes.io/projected/5ee5efe1-368e-476e-8516-d9b81d9a38a6-kube-api-access-h8hm5\") pod \"nova-operator-controller-manager-74b6b5dc96-m5zqv\" (UID: \"5ee5efe1-368e-476e-8516-d9b81d9a38a6\") " pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-m5zqv" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.105532 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkbwh\" (UniqueName: \"kubernetes.io/projected/2192fb73-bd5d-440f-8c01-2871658164d9-kube-api-access-rkbwh\") pod \"ironic-operator-controller-manager-545456dc4-fp474\" (UID: \"2192fb73-bd5d-440f-8c01-2871658164d9\") " pod="openstack-operators/ironic-operator-controller-manager-545456dc4-fp474" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.105558 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz9lb\" (UniqueName: \"kubernetes.io/projected/9b633770-3463-40ee-b5fc-48ec5fb4846c-kube-api-access-wz9lb\") pod \"keystone-operator-controller-manager-55ffd4876b-nxrtn\" (UID: \"9b633770-3463-40ee-b5fc-48ec5fb4846c\") " pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-nxrtn" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.105583 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5n9s\" (UniqueName: \"kubernetes.io/projected/2d989d20-0a79-448b-8b73-ad9378119232-kube-api-access-t5n9s\") pod \"infra-operator-controller-manager-f7fcc58b9-qc48n\" (UID: \"2d989d20-0a79-448b-8b73-ad9378119232\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.105631 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tw5d\" (UniqueName: \"kubernetes.io/projected/71ebdc25-566e-4b42-8f48-6f3c52e1125c-kube-api-access-7tw5d\") pod \"mariadb-operator-controller-manager-556b8b874-f4dv9\" (UID: \"71ebdc25-566e-4b42-8f48-6f3c52e1125c\") " pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-f4dv9" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.105666 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4whc\" (UniqueName: \"kubernetes.io/projected/50d18356-e250-46b6-bcd3-053f770f8f58-kube-api-access-k4whc\") pod \"neutron-operator-controller-manager-54688575f-q2gkq\" (UID: \"50d18356-e250-46b6-bcd3-053f770f8f58\") " pod="openstack-operators/neutron-operator-controller-manager-54688575f-q2gkq" Feb 27 16:40:37 crc kubenswrapper[4814]: E0227 16:40:37.105845 4814 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 27 16:40:37 crc kubenswrapper[4814]: E0227 16:40:37.105899 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert podName:2d989d20-0a79-448b-8b73-ad9378119232 nodeName:}" failed. No retries permitted until 2026-02-27 16:40:37.605878982 +0000 UTC m=+1050.058503822 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert") pod "infra-operator-controller-manager-f7fcc58b9-qc48n" (UID: "2d989d20-0a79-448b-8b73-ad9378119232") : secret "infra-operator-webhook-server-cert" not found Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.106149 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9b9ff9f4d-pq6mn"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.106930 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-pq6mn" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.110770 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-648564c9fc-hbdxk"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.111924 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-w9gnk" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.112067 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-bgp96" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.116293 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9b9ff9f4d-pq6mn"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.126422 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-9kqgh" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.167517 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkbwh\" (UniqueName: \"kubernetes.io/projected/2192fb73-bd5d-440f-8c01-2871658164d9-kube-api-access-rkbwh\") pod \"ironic-operator-controller-manager-545456dc4-fp474\" (UID: \"2192fb73-bd5d-440f-8c01-2871658164d9\") " pod="openstack-operators/ironic-operator-controller-manager-545456dc4-fp474" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.169837 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5n9s\" (UniqueName: \"kubernetes.io/projected/2d989d20-0a79-448b-8b73-ad9378119232-kube-api-access-t5n9s\") pod \"infra-operator-controller-manager-f7fcc58b9-qc48n\" (UID: \"2d989d20-0a79-448b-8b73-ad9378119232\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.171490 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5fdb694969-4s2gh"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.175901 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz9lb\" (UniqueName: \"kubernetes.io/projected/9b633770-3463-40ee-b5fc-48ec5fb4846c-kube-api-access-wz9lb\") pod \"keystone-operator-controller-manager-55ffd4876b-nxrtn\" (UID: \"9b633770-3463-40ee-b5fc-48ec5fb4846c\") " pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-nxrtn" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.181836 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-ccrlj" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.183676 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-4s2gh" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.187965 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-m5ld6" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.191309 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5fdb694969-4s2gh"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.197895 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-p74f5" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.207509 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tw5d\" (UniqueName: \"kubernetes.io/projected/71ebdc25-566e-4b42-8f48-6f3c52e1125c-kube-api-access-7tw5d\") pod \"mariadb-operator-controller-manager-556b8b874-f4dv9\" (UID: \"71ebdc25-566e-4b42-8f48-6f3c52e1125c\") " pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-f4dv9" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.207557 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4whc\" (UniqueName: \"kubernetes.io/projected/50d18356-e250-46b6-bcd3-053f770f8f58-kube-api-access-k4whc\") pod \"neutron-operator-controller-manager-54688575f-q2gkq\" (UID: \"50d18356-e250-46b6-bcd3-053f770f8f58\") " pod="openstack-operators/neutron-operator-controller-manager-54688575f-q2gkq" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.207594 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfwc7\" (UniqueName: \"kubernetes.io/projected/103a95f2-68f7-41b0-9dea-46cde1679842-kube-api-access-jfwc7\") pod \"placement-operator-controller-manager-648564c9fc-hbdxk\" (UID: \"103a95f2-68f7-41b0-9dea-46cde1679842\") " pod="openstack-operators/placement-operator-controller-manager-648564c9fc-hbdxk" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.207620 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwttl\" (UniqueName: \"kubernetes.io/projected/edf51635-71a3-4c03-8aef-50212a3b8247-kube-api-access-lwttl\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr\" (UID: \"edf51635-71a3-4c03-8aef-50212a3b8247\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.207643 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8cdq\" (UniqueName: \"kubernetes.io/projected/dfcfffcb-36da-4853-8ed7-ba76caba9d3c-kube-api-access-s8cdq\") pod \"manila-operator-controller-manager-67d996989d-c5mcj\" (UID: \"dfcfffcb-36da-4853-8ed7-ba76caba9d3c\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-c5mcj" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.207669 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh4qv\" (UniqueName: \"kubernetes.io/projected/c6faf190-d834-45e6-8a18-067067c0bef5-kube-api-access-kh4qv\") pod \"ovn-operator-controller-manager-75684d597f-5958h\" (UID: \"c6faf190-d834-45e6-8a18-067067c0bef5\") " pod="openstack-operators/ovn-operator-controller-manager-75684d597f-5958h" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.207699 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qszk7\" (UniqueName: \"kubernetes.io/projected/75ee2f25-f59a-4a60-8245-1b638957d234-kube-api-access-qszk7\") pod \"octavia-operator-controller-manager-5d86c7ddb7-4m2wv\" (UID: \"75ee2f25-f59a-4a60-8245-1b638957d234\") " pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-4m2wv" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.207719 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8hm5\" (UniqueName: \"kubernetes.io/projected/5ee5efe1-368e-476e-8516-d9b81d9a38a6-kube-api-access-h8hm5\") pod \"nova-operator-controller-manager-74b6b5dc96-m5zqv\" (UID: \"5ee5efe1-368e-476e-8516-d9b81d9a38a6\") " pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-m5zqv" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.207739 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s26fx\" (UniqueName: \"kubernetes.io/projected/59cb5f17-969c-4c4f-81bf-a5b0e65acace-kube-api-access-s26fx\") pod \"swift-operator-controller-manager-9b9ff9f4d-pq6mn\" (UID: \"59cb5f17-969c-4c4f-81bf-a5b0e65acace\") " pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-pq6mn" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.207765 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr\" (UID: \"edf51635-71a3-4c03-8aef-50212a3b8247\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.226414 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-55b5ff4dbb-tck96"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.227588 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-tck96" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.234631 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-8kghv" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.236931 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8cdq\" (UniqueName: \"kubernetes.io/projected/dfcfffcb-36da-4853-8ed7-ba76caba9d3c-kube-api-access-s8cdq\") pod \"manila-operator-controller-manager-67d996989d-c5mcj\" (UID: \"dfcfffcb-36da-4853-8ed7-ba76caba9d3c\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-c5mcj" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.242859 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qszk7\" (UniqueName: \"kubernetes.io/projected/75ee2f25-f59a-4a60-8245-1b638957d234-kube-api-access-qszk7\") pod \"octavia-operator-controller-manager-5d86c7ddb7-4m2wv\" (UID: \"75ee2f25-f59a-4a60-8245-1b638957d234\") " pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-4m2wv" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.248205 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tw5d\" (UniqueName: \"kubernetes.io/projected/71ebdc25-566e-4b42-8f48-6f3c52e1125c-kube-api-access-7tw5d\") pod \"mariadb-operator-controller-manager-556b8b874-f4dv9\" (UID: \"71ebdc25-566e-4b42-8f48-6f3c52e1125c\") " pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-f4dv9" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.248297 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8hm5\" (UniqueName: \"kubernetes.io/projected/5ee5efe1-368e-476e-8516-d9b81d9a38a6-kube-api-access-h8hm5\") pod \"nova-operator-controller-manager-74b6b5dc96-m5zqv\" (UID: \"5ee5efe1-368e-476e-8516-d9b81d9a38a6\") " pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-m5zqv" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.253892 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4whc\" (UniqueName: \"kubernetes.io/projected/50d18356-e250-46b6-bcd3-053f770f8f58-kube-api-access-k4whc\") pod \"neutron-operator-controller-manager-54688575f-q2gkq\" (UID: \"50d18356-e250-46b6-bcd3-053f770f8f58\") " pod="openstack-operators/neutron-operator-controller-manager-54688575f-q2gkq" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.260646 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-55b5ff4dbb-tck96"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.265584 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-54688575f-q2gkq" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.275211 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-m5zqv" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.289077 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-4m2wv" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.311316 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh4qv\" (UniqueName: \"kubernetes.io/projected/c6faf190-d834-45e6-8a18-067067c0bef5-kube-api-access-kh4qv\") pod \"ovn-operator-controller-manager-75684d597f-5958h\" (UID: \"c6faf190-d834-45e6-8a18-067067c0bef5\") " pod="openstack-operators/ovn-operator-controller-manager-75684d597f-5958h" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.311390 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s26fx\" (UniqueName: \"kubernetes.io/projected/59cb5f17-969c-4c4f-81bf-a5b0e65acace-kube-api-access-s26fx\") pod \"swift-operator-controller-manager-9b9ff9f4d-pq6mn\" (UID: \"59cb5f17-969c-4c4f-81bf-a5b0e65acace\") " pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-pq6mn" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.311422 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-785cv\" (UniqueName: \"kubernetes.io/projected/3f85fa1d-a7c6-49b8-8f85-d02670fa2849-kube-api-access-785cv\") pod \"telemetry-operator-controller-manager-5fdb694969-4s2gh\" (UID: \"3f85fa1d-a7c6-49b8-8f85-d02670fa2849\") " pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-4s2gh" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.311444 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr\" (UID: \"edf51635-71a3-4c03-8aef-50212a3b8247\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.311533 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfwc7\" (UniqueName: \"kubernetes.io/projected/103a95f2-68f7-41b0-9dea-46cde1679842-kube-api-access-jfwc7\") pod \"placement-operator-controller-manager-648564c9fc-hbdxk\" (UID: \"103a95f2-68f7-41b0-9dea-46cde1679842\") " pod="openstack-operators/placement-operator-controller-manager-648564c9fc-hbdxk" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.311554 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwttl\" (UniqueName: \"kubernetes.io/projected/edf51635-71a3-4c03-8aef-50212a3b8247-kube-api-access-lwttl\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr\" (UID: \"edf51635-71a3-4c03-8aef-50212a3b8247\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.311570 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvzk4\" (UniqueName: \"kubernetes.io/projected/30b75b23-8950-40e1-a515-eaf136867a25-kube-api-access-hvzk4\") pod \"test-operator-controller-manager-55b5ff4dbb-tck96\" (UID: \"30b75b23-8950-40e1-a515-eaf136867a25\") " pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-tck96" Feb 27 16:40:37 crc kubenswrapper[4814]: E0227 16:40:37.312909 4814 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 16:40:37 crc kubenswrapper[4814]: E0227 16:40:37.312955 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert podName:edf51635-71a3-4c03-8aef-50212a3b8247 nodeName:}" failed. No retries permitted until 2026-02-27 16:40:37.812940015 +0000 UTC m=+1050.265564835 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" (UID: "edf51635-71a3-4c03-8aef-50212a3b8247") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.329689 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-bccc79885-hh455"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.331695 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-hh455" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.337562 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfwc7\" (UniqueName: \"kubernetes.io/projected/103a95f2-68f7-41b0-9dea-46cde1679842-kube-api-access-jfwc7\") pod \"placement-operator-controller-manager-648564c9fc-hbdxk\" (UID: \"103a95f2-68f7-41b0-9dea-46cde1679842\") " pod="openstack-operators/placement-operator-controller-manager-648564c9fc-hbdxk" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.338366 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-dwjqj" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.340668 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-bccc79885-hh455"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.347679 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwttl\" (UniqueName: \"kubernetes.io/projected/edf51635-71a3-4c03-8aef-50212a3b8247-kube-api-access-lwttl\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr\" (UID: \"edf51635-71a3-4c03-8aef-50212a3b8247\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.355518 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s26fx\" (UniqueName: \"kubernetes.io/projected/59cb5f17-969c-4c4f-81bf-a5b0e65acace-kube-api-access-s26fx\") pod \"swift-operator-controller-manager-9b9ff9f4d-pq6mn\" (UID: \"59cb5f17-969c-4c4f-81bf-a5b0e65acace\") " pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-pq6mn" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.371396 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh4qv\" (UniqueName: \"kubernetes.io/projected/c6faf190-d834-45e6-8a18-067067c0bef5-kube-api-access-kh4qv\") pod \"ovn-operator-controller-manager-75684d597f-5958h\" (UID: \"c6faf190-d834-45e6-8a18-067067c0bef5\") " pod="openstack-operators/ovn-operator-controller-manager-75684d597f-5958h" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.384415 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-5958h" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.393861 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-hbdxk" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.412420 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-785cv\" (UniqueName: \"kubernetes.io/projected/3f85fa1d-a7c6-49b8-8f85-d02670fa2849-kube-api-access-785cv\") pod \"telemetry-operator-controller-manager-5fdb694969-4s2gh\" (UID: \"3f85fa1d-a7c6-49b8-8f85-d02670fa2849\") " pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-4s2gh" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.412713 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slmn8\" (UniqueName: \"kubernetes.io/projected/4b02bae2-c5ac-423e-a2f1-b95de1264f47-kube-api-access-slmn8\") pod \"watcher-operator-controller-manager-bccc79885-hh455\" (UID: \"4b02bae2-c5ac-423e-a2f1-b95de1264f47\") " pod="openstack-operators/watcher-operator-controller-manager-bccc79885-hh455" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.412779 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvzk4\" (UniqueName: \"kubernetes.io/projected/30b75b23-8950-40e1-a515-eaf136867a25-kube-api-access-hvzk4\") pod \"test-operator-controller-manager-55b5ff4dbb-tck96\" (UID: \"30b75b23-8950-40e1-a515-eaf136867a25\") " pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-tck96" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.428282 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-fp474" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.431819 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-nxrtn" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.438052 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.438959 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.441598 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.441727 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.441765 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-xjqww" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.442962 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-785cv\" (UniqueName: \"kubernetes.io/projected/3f85fa1d-a7c6-49b8-8f85-d02670fa2849-kube-api-access-785cv\") pod \"telemetry-operator-controller-manager-5fdb694969-4s2gh\" (UID: \"3f85fa1d-a7c6-49b8-8f85-d02670fa2849\") " pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-4s2gh" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.448177 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.449358 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvzk4\" (UniqueName: \"kubernetes.io/projected/30b75b23-8950-40e1-a515-eaf136867a25-kube-api-access-hvzk4\") pod \"test-operator-controller-manager-55b5ff4dbb-tck96\" (UID: \"30b75b23-8950-40e1-a515-eaf136867a25\") " pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-tck96" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.475351 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-pq6mn" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.481168 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-67d996989d-c5mcj" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.495831 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-4s2gh" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.515371 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.515451 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxdw8\" (UniqueName: \"kubernetes.io/projected/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-kube-api-access-bxdw8\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.515483 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.515559 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slmn8\" (UniqueName: \"kubernetes.io/projected/4b02bae2-c5ac-423e-a2f1-b95de1264f47-kube-api-access-slmn8\") pod \"watcher-operator-controller-manager-bccc79885-hh455\" (UID: \"4b02bae2-c5ac-423e-a2f1-b95de1264f47\") " pod="openstack-operators/watcher-operator-controller-manager-bccc79885-hh455" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.518553 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-tck96" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.519324 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-f4dv9" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.539437 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwz52"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.540391 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwz52" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.547043 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-qjg2b" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.558390 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwz52"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.566176 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slmn8\" (UniqueName: \"kubernetes.io/projected/4b02bae2-c5ac-423e-a2f1-b95de1264f47-kube-api-access-slmn8\") pod \"watcher-operator-controller-manager-bccc79885-hh455\" (UID: \"4b02bae2-c5ac-423e-a2f1-b95de1264f47\") " pod="openstack-operators/watcher-operator-controller-manager-bccc79885-hh455" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.617064 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxdw8\" (UniqueName: \"kubernetes.io/projected/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-kube-api-access-bxdw8\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.617117 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.617144 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-qc48n\" (UID: \"2d989d20-0a79-448b-8b73-ad9378119232\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.617300 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.617326 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv2kk\" (UniqueName: \"kubernetes.io/projected/60ad038c-aad7-4309-89cf-17bf1bda00ac-kube-api-access-zv2kk\") pod \"rabbitmq-cluster-operator-manager-668c99d594-lwz52\" (UID: \"60ad038c-aad7-4309-89cf-17bf1bda00ac\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwz52" Feb 27 16:40:37 crc kubenswrapper[4814]: E0227 16:40:37.618927 4814 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 16:40:37 crc kubenswrapper[4814]: E0227 16:40:37.618964 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs podName:96d3c178-c0f0-4bd5-b3bd-0e572404d30a nodeName:}" failed. No retries permitted until 2026-02-27 16:40:38.118950243 +0000 UTC m=+1050.571575073 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs") pod "openstack-operator-controller-manager-655dd5d898-d7f5g" (UID: "96d3c178-c0f0-4bd5-b3bd-0e572404d30a") : secret "webhook-server-cert" not found Feb 27 16:40:37 crc kubenswrapper[4814]: E0227 16:40:37.619082 4814 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 27 16:40:37 crc kubenswrapper[4814]: E0227 16:40:37.619102 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert podName:2d989d20-0a79-448b-8b73-ad9378119232 nodeName:}" failed. No retries permitted until 2026-02-27 16:40:38.619095427 +0000 UTC m=+1051.071720247 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert") pod "infra-operator-controller-manager-f7fcc58b9-qc48n" (UID: "2d989d20-0a79-448b-8b73-ad9378119232") : secret "infra-operator-webhook-server-cert" not found Feb 27 16:40:37 crc kubenswrapper[4814]: E0227 16:40:37.620436 4814 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 16:40:37 crc kubenswrapper[4814]: E0227 16:40:37.620465 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs podName:96d3c178-c0f0-4bd5-b3bd-0e572404d30a nodeName:}" failed. No retries permitted until 2026-02-27 16:40:38.120457859 +0000 UTC m=+1050.573082689 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs") pod "openstack-operator-controller-manager-655dd5d898-d7f5g" (UID: "96d3c178-c0f0-4bd5-b3bd-0e572404d30a") : secret "metrics-server-cert" not found Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.646399 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-2d2n7"] Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.652892 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxdw8\" (UniqueName: \"kubernetes.io/projected/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-kube-api-access-bxdw8\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.719879 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv2kk\" (UniqueName: \"kubernetes.io/projected/60ad038c-aad7-4309-89cf-17bf1bda00ac-kube-api-access-zv2kk\") pod \"rabbitmq-cluster-operator-manager-668c99d594-lwz52\" (UID: \"60ad038c-aad7-4309-89cf-17bf1bda00ac\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwz52" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.754736 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv2kk\" (UniqueName: \"kubernetes.io/projected/60ad038c-aad7-4309-89cf-17bf1bda00ac-kube-api-access-zv2kk\") pod \"rabbitmq-cluster-operator-manager-668c99d594-lwz52\" (UID: \"60ad038c-aad7-4309-89cf-17bf1bda00ac\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwz52" Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.760467 4814 generic.go:334] "Generic (PLEG): container finished" podID="afffcacd-d821-4cfd-aad8-462aafd2986f" containerID="73a1b5d0586039162b6eb935fdfed5ecf01f42d81a7bd43b2f1bf99c52420a28" exitCode=0 Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.760513 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nvzj5" event={"ID":"afffcacd-d821-4cfd-aad8-462aafd2986f","Type":"ContainerDied","Data":"73a1b5d0586039162b6eb935fdfed5ecf01f42d81a7bd43b2f1bf99c52420a28"} Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.821214 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr\" (UID: \"edf51635-71a3-4c03-8aef-50212a3b8247\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" Feb 27 16:40:37 crc kubenswrapper[4814]: E0227 16:40:37.821365 4814 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 16:40:37 crc kubenswrapper[4814]: E0227 16:40:37.821410 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert podName:edf51635-71a3-4c03-8aef-50212a3b8247 nodeName:}" failed. No retries permitted until 2026-02-27 16:40:38.821397285 +0000 UTC m=+1051.274022115 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" (UID: "edf51635-71a3-4c03-8aef-50212a3b8247") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 16:40:37 crc kubenswrapper[4814]: I0227 16:40:37.866315 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-hh455" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.025538 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwz52" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.125984 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.126056 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.126185 4814 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.126231 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs podName:96d3c178-c0f0-4bd5-b3bd-0e572404d30a nodeName:}" failed. No retries permitted until 2026-02-27 16:40:39.126217717 +0000 UTC m=+1051.578842547 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs") pod "openstack-operator-controller-manager-655dd5d898-d7f5g" (UID: "96d3c178-c0f0-4bd5-b3bd-0e572404d30a") : secret "webhook-server-cert" not found Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.126560 4814 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.126586 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs podName:96d3c178-c0f0-4bd5-b3bd-0e572404d30a nodeName:}" failed. No retries permitted until 2026-02-27 16:40:39.126579468 +0000 UTC m=+1051.579204288 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs") pod "openstack-operator-controller-manager-655dd5d898-d7f5g" (UID: "96d3c178-c0f0-4bd5-b3bd-0e572404d30a") : secret "metrics-server-cert" not found Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.310352 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-5d87c9d997-pjncz"] Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.332787 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-64db6967f8-9kqgh"] Feb 27 16:40:38 crc kubenswrapper[4814]: W0227 16:40:38.337456 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49b03138_9b4c_486c_af91_37c16e8c2536.slice/crio-75e69669cf97061f9ac6d2bd2bf587de64aed4af2c0e717131185a85c184f7e9 WatchSource:0}: Error finding container 75e69669cf97061f9ac6d2bd2bf587de64aed4af2c0e717131185a85c184f7e9: Status 404 returned error can't find the container with id 75e69669cf97061f9ac6d2bd2bf587de64aed4af2c0e717131185a85c184f7e9 Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.357634 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.415659 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6db6876945-xnhhg"] Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.429729 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5rjt\" (UniqueName: \"kubernetes.io/projected/afffcacd-d821-4cfd-aad8-462aafd2986f-kube-api-access-s5rjt\") pod \"afffcacd-d821-4cfd-aad8-462aafd2986f\" (UID: \"afffcacd-d821-4cfd-aad8-462aafd2986f\") " Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.429900 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afffcacd-d821-4cfd-aad8-462aafd2986f-catalog-content\") pod \"afffcacd-d821-4cfd-aad8-462aafd2986f\" (UID: \"afffcacd-d821-4cfd-aad8-462aafd2986f\") " Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.429927 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afffcacd-d821-4cfd-aad8-462aafd2986f-utilities\") pod \"afffcacd-d821-4cfd-aad8-462aafd2986f\" (UID: \"afffcacd-d821-4cfd-aad8-462aafd2986f\") " Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.430886 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afffcacd-d821-4cfd-aad8-462aafd2986f-utilities" (OuterVolumeSpecName: "utilities") pod "afffcacd-d821-4cfd-aad8-462aafd2986f" (UID: "afffcacd-d821-4cfd-aad8-462aafd2986f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.438171 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afffcacd-d821-4cfd-aad8-462aafd2986f-kube-api-access-s5rjt" (OuterVolumeSpecName: "kube-api-access-s5rjt") pod "afffcacd-d821-4cfd-aad8-462aafd2986f" (UID: "afffcacd-d821-4cfd-aad8-462aafd2986f"). InnerVolumeSpecName "kube-api-access-s5rjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.499999 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afffcacd-d821-4cfd-aad8-462aafd2986f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "afffcacd-d821-4cfd-aad8-462aafd2986f" (UID: "afffcacd-d821-4cfd-aad8-462aafd2986f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.539885 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afffcacd-d821-4cfd-aad8-462aafd2986f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.539917 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afffcacd-d821-4cfd-aad8-462aafd2986f-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.539926 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5rjt\" (UniqueName: \"kubernetes.io/projected/afffcacd-d821-4cfd-aad8-462aafd2986f-kube-api-access-s5rjt\") on node \"crc\" DevicePath \"\"" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.641540 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-qc48n\" (UID: \"2d989d20-0a79-448b-8b73-ad9378119232\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.641731 4814 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.641811 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert podName:2d989d20-0a79-448b-8b73-ad9378119232 nodeName:}" failed. No retries permitted until 2026-02-27 16:40:40.641791073 +0000 UTC m=+1053.094415903 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert") pod "infra-operator-controller-manager-f7fcc58b9-qc48n" (UID: "2d989d20-0a79-448b-8b73-ad9378119232") : secret "infra-operator-webhook-server-cert" not found Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.734366 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-545456dc4-fp474"] Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.750123 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54688575f-q2gkq"] Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.757002 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-75684d597f-5958h"] Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.766799 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-74b6b5dc96-m5zqv"] Feb 27 16:40:38 crc kubenswrapper[4814]: W0227 16:40:38.772148 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod103a95f2_68f7_41b0_9dea_46cde1679842.slice/crio-e73139de008ef4fecc392e1d1e1bc9f8fa39a0f29e6d63fe21f25e6bc752a452 WatchSource:0}: Error finding container e73139de008ef4fecc392e1d1e1bc9f8fa39a0f29e6d63fe21f25e6bc752a452: Status 404 returned error can't find the container with id e73139de008ef4fecc392e1d1e1bc9f8fa39a0f29e6d63fe21f25e6bc752a452 Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.772445 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-648564c9fc-hbdxk"] Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.782461 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54688575f-q2gkq" event={"ID":"50d18356-e250-46b6-bcd3-053f770f8f58","Type":"ContainerStarted","Data":"40cbaa2c1c1c2408816a710c317659c5af4f10070cf7a74bb85fb5b989a03891"} Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.784628 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55ffd4876b-nxrtn"] Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.789926 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-p74f5"] Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.794450 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-c5mcj"] Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.797978 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nvzj5" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.797976 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nvzj5" event={"ID":"afffcacd-d821-4cfd-aad8-462aafd2986f","Type":"ContainerDied","Data":"de4ed563d472ca6c817f216145bff88ec6d1c30ea8c10eb398e463dbebd74197"} Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.798036 4814 scope.go:117] "RemoveContainer" containerID="73a1b5d0586039162b6eb935fdfed5ecf01f42d81a7bd43b2f1bf99c52420a28" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.803491 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-m5zqv" event={"ID":"5ee5efe1-368e-476e-8516-d9b81d9a38a6","Type":"ContainerStarted","Data":"0896835bdd1384cd26bda72bbdb21029f013341d4392ed5be9d818ce11c969fd"} Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.805303 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-4m2wv"] Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.809900 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-2d2n7" event={"ID":"de6e87db-4283-4aab-a97e-23792fd3b480","Type":"ContainerStarted","Data":"f090f47f5e92defb49ed0652bcea73bf4e2845bf4c6d7f071b669f923e3f0e7c"} Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.811174 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-cf99c678f-ccrlj"] Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.813364 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-9kqgh" event={"ID":"49b03138-9b4c-486c-af91-37c16e8c2536","Type":"ContainerStarted","Data":"75e69669cf97061f9ac6d2bd2bf587de64aed4af2c0e717131185a85c184f7e9"} Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.816685 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:114c0dee0bab1d453890e9dcc7727de749055bdbea049384d5696e7ac8d78fe3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-46gnd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-78bc7f9bd9-p74f5_openstack-operators(ea6ecef0-2796-4c32-9a9a-ee178574d2f9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.816832 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5fdb694969-4s2gh"] Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.818733 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-xnhhg" event={"ID":"6ede514a-7a77-4cfe-a746-3d7bd2f5d7fd","Type":"ContainerStarted","Data":"8cba92a8562fc92efcc74bf8e745079df182479cd757e7bdd168c8ce1ab0a9a0"} Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.821235 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-p74f5" podUID="ea6ecef0-2796-4c32-9a9a-ee178574d2f9" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.824433 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9b9ff9f4d-pq6mn"] Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.824563 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-pjncz" event={"ID":"6ee0f0c6-52ee-4d20-999f-a0a579dcdaef","Type":"ContainerStarted","Data":"3c5357781180ae8499c9cb43f6f6e65d2c528a76290d80a742ea22fdc4e8f41e"} Feb 27 16:40:38 crc kubenswrapper[4814]: W0227 16:40:38.828645 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f85fa1d_a7c6_49b8_8f85_d02670fa2849.slice/crio-e9dd0242c4fcdd238e4d612c3b39a299105ade98e3a126700b31c7e84b994958 WatchSource:0}: Error finding container e9dd0242c4fcdd238e4d612c3b39a299105ade98e3a126700b31c7e84b994958: Status 404 returned error can't find the container with id e9dd0242c4fcdd238e4d612c3b39a299105ade98e3a126700b31c7e84b994958 Feb 27 16:40:38 crc kubenswrapper[4814]: W0227 16:40:38.830513 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30b75b23_8950_40e1_a515_eaf136867a25.slice/crio-1fee8ec8e70e77c7ff76a8ad1ad36d25c3fdcfe88410d6224f7cb7c69f86519f WatchSource:0}: Error finding container 1fee8ec8e70e77c7ff76a8ad1ad36d25c3fdcfe88410d6224f7cb7c69f86519f: Status 404 returned error can't find the container with id 1fee8ec8e70e77c7ff76a8ad1ad36d25c3fdcfe88410d6224f7cb7c69f86519f Feb 27 16:40:38 crc kubenswrapper[4814]: W0227 16:40:38.836432 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60ad038c_aad7_4309_89cf_17bf1bda00ac.slice/crio-2b38ea9b0aebf370fc1b10736704c549f759ca2c120a1d6f6eaf04c69483b28f WatchSource:0}: Error finding container 2b38ea9b0aebf370fc1b10736704c549f759ca2c120a1d6f6eaf04c69483b28f: Status 404 returned error can't find the container with id 2b38ea9b0aebf370fc1b10736704c549f759ca2c120a1d6f6eaf04c69483b28f Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.836515 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-55b5ff4dbb-tck96"] Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.842811 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-556b8b874-f4dv9"] Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.843690 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr\" (UID: \"edf51635-71a3-4c03-8aef-50212a3b8247\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.843857 4814 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.843902 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert podName:edf51635-71a3-4c03-8aef-50212a3b8247 nodeName:}" failed. No retries permitted until 2026-02-27 16:40:40.843887064 +0000 UTC m=+1053.296511894 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" (UID: "edf51635-71a3-4c03-8aef-50212a3b8247") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.845610 4814 scope.go:117] "RemoveContainer" containerID="9d3f7a045fdbb41ede0f5050fc4f55a92ead8ca0c72ec9718d0151a396028245" Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.845859 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zv2kk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-lwz52_openstack-operators(60ad038c-aad7-4309-89cf-17bf1bda00ac): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.845901 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:1b9074a4ce16396d8bd2d30a475fc8c2f004f75a023e3eef8950661e89c0bcc6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-785cv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5fdb694969-4s2gh_openstack-operators(3f85fa1d-a7c6-49b8-8f85-d02670fa2849): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.846011 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:9d03f03aa9a460f1fcac8875064808c03e4ecd0388873bbfb9c7dc58331f3968,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hvzk4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-55b5ff4dbb-tck96_openstack-operators(30b75b23-8950-40e1-a515-eaf136867a25): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.847135 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-tck96" podUID="30b75b23-8950-40e1-a515-eaf136867a25" Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.847398 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-4s2gh" podUID="3f85fa1d-a7c6-49b8-8f85-d02670fa2849" Feb 27 16:40:38 crc kubenswrapper[4814]: W0227 16:40:38.847760 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59cb5f17_969c_4c4f_81bf_a5b0e65acace.slice/crio-188337e8225fb70d2aed8f5f59b44937ec86cca84f9a9d91ce5e52e665ee60e5 WatchSource:0}: Error finding container 188337e8225fb70d2aed8f5f59b44937ec86cca84f9a9d91ce5e52e665ee60e5: Status 404 returned error can't find the container with id 188337e8225fb70d2aed8f5f59b44937ec86cca84f9a9d91ce5e52e665ee60e5 Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.847835 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwz52" podUID="60ad038c-aad7-4309-89cf-17bf1bda00ac" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.849000 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwz52"] Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.857558 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:71f2ab3bb41d1743287a3270dd49e32192b347d8ba7353d2250cbd7e8528219b,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7tw5d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-556b8b874-f4dv9_openstack-operators(71ebdc25-566e-4b42-8f48-6f3c52e1125c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.857679 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:f309cdea8084a4b1e8cbcd732d6e250fd93c55cfd1b48ba9026907c8591faab7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s26fx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9b9ff9f4d-pq6mn_openstack-operators(59cb5f17-969c-4c4f-81bf-a5b0e65acace): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.857748 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nvzj5"] Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.858721 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-f4dv9" podUID="71ebdc25-566e-4b42-8f48-6f3c52e1125c" Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.858772 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-pq6mn" podUID="59cb5f17-969c-4c4f-81bf-a5b0e65acace" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.862374 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-bccc79885-hh455"] Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.870091 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nvzj5"] Feb 27 16:40:38 crc kubenswrapper[4814]: W0227 16:40:38.878887 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b02bae2_c5ac_423e_a2f1_b95de1264f47.slice/crio-ab0ee9c0aba5a7fcbb60238d6bb3d99de4aa8f9aa3aa840402510f146e0fe07a WatchSource:0}: Error finding container ab0ee9c0aba5a7fcbb60238d6bb3d99de4aa8f9aa3aa840402510f146e0fe07a: Status 404 returned error can't find the container with id ab0ee9c0aba5a7fcbb60238d6bb3d99de4aa8f9aa3aa840402510f146e0fe07a Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.882281 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:06311600a491c689493552e7ff26e36df740fa4e7c143fca874bef19f24afb97,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-slmn8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-bccc79885-hh455_openstack-operators(4b02bae2-c5ac-423e-a2f1-b95de1264f47): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 16:40:38 crc kubenswrapper[4814]: E0227 16:40:38.883484 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-hh455" podUID="4b02bae2-c5ac-423e-a2f1-b95de1264f47" Feb 27 16:40:38 crc kubenswrapper[4814]: I0227 16:40:38.920440 4814 scope.go:117] "RemoveContainer" containerID="2aee2051d874c3163014ff09ec4d518d320944c0447ff89117be8323300176e5" Feb 27 16:40:39 crc kubenswrapper[4814]: I0227 16:40:39.148340 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:39 crc kubenswrapper[4814]: I0227 16:40:39.148436 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:39 crc kubenswrapper[4814]: E0227 16:40:39.148615 4814 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 16:40:39 crc kubenswrapper[4814]: E0227 16:40:39.148633 4814 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 16:40:39 crc kubenswrapper[4814]: E0227 16:40:39.148677 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs podName:96d3c178-c0f0-4bd5-b3bd-0e572404d30a nodeName:}" failed. No retries permitted until 2026-02-27 16:40:41.148661655 +0000 UTC m=+1053.601286485 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs") pod "openstack-operator-controller-manager-655dd5d898-d7f5g" (UID: "96d3c178-c0f0-4bd5-b3bd-0e572404d30a") : secret "webhook-server-cert" not found Feb 27 16:40:39 crc kubenswrapper[4814]: E0227 16:40:39.148727 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs podName:96d3c178-c0f0-4bd5-b3bd-0e572404d30a nodeName:}" failed. No retries permitted until 2026-02-27 16:40:41.148701936 +0000 UTC m=+1053.601326856 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs") pod "openstack-operator-controller-manager-655dd5d898-d7f5g" (UID: "96d3c178-c0f0-4bd5-b3bd-0e572404d30a") : secret "metrics-server-cert" not found Feb 27 16:40:39 crc kubenswrapper[4814]: I0227 16:40:39.859445 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-fp474" event={"ID":"2192fb73-bd5d-440f-8c01-2871658164d9","Type":"ContainerStarted","Data":"c82d7d70613b5bf5d5f663ae69a5036787a4926dcbee2cc8abeac2803bf637e5"} Feb 27 16:40:39 crc kubenswrapper[4814]: I0227 16:40:39.875041 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwz52" event={"ID":"60ad038c-aad7-4309-89cf-17bf1bda00ac","Type":"ContainerStarted","Data":"2b38ea9b0aebf370fc1b10736704c549f759ca2c120a1d6f6eaf04c69483b28f"} Feb 27 16:40:39 crc kubenswrapper[4814]: I0227 16:40:39.878157 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-pq6mn" event={"ID":"59cb5f17-969c-4c4f-81bf-a5b0e65acace","Type":"ContainerStarted","Data":"188337e8225fb70d2aed8f5f59b44937ec86cca84f9a9d91ce5e52e665ee60e5"} Feb 27 16:40:39 crc kubenswrapper[4814]: E0227 16:40:39.879896 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:f309cdea8084a4b1e8cbcd732d6e250fd93c55cfd1b48ba9026907c8591faab7\\\"\"" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-pq6mn" podUID="59cb5f17-969c-4c4f-81bf-a5b0e65acace" Feb 27 16:40:39 crc kubenswrapper[4814]: I0227 16:40:39.881047 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-4s2gh" event={"ID":"3f85fa1d-a7c6-49b8-8f85-d02670fa2849","Type":"ContainerStarted","Data":"e9dd0242c4fcdd238e4d612c3b39a299105ade98e3a126700b31c7e84b994958"} Feb 27 16:40:39 crc kubenswrapper[4814]: E0227 16:40:39.887144 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:1b9074a4ce16396d8bd2d30a475fc8c2f004f75a023e3eef8950661e89c0bcc6\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-4s2gh" podUID="3f85fa1d-a7c6-49b8-8f85-d02670fa2849" Feb 27 16:40:39 crc kubenswrapper[4814]: I0227 16:40:39.887309 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-nxrtn" event={"ID":"9b633770-3463-40ee-b5fc-48ec5fb4846c","Type":"ContainerStarted","Data":"89fae0165a4c02dfc44551ddaf42c592fd2a0388769b8e1774fb3d2b673a04f6"} Feb 27 16:40:39 crc kubenswrapper[4814]: E0227 16:40:39.904226 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwz52" podUID="60ad038c-aad7-4309-89cf-17bf1bda00ac" Feb 27 16:40:39 crc kubenswrapper[4814]: I0227 16:40:39.913129 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-p74f5" event={"ID":"ea6ecef0-2796-4c32-9a9a-ee178574d2f9","Type":"ContainerStarted","Data":"b276b67c6c490918ef4660c687fc197541b4295f064f82fff29b41bc43464cbc"} Feb 27 16:40:39 crc kubenswrapper[4814]: E0227 16:40:39.915077 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:114c0dee0bab1d453890e9dcc7727de749055bdbea049384d5696e7ac8d78fe3\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-p74f5" podUID="ea6ecef0-2796-4c32-9a9a-ee178574d2f9" Feb 27 16:40:39 crc kubenswrapper[4814]: I0227 16:40:39.916054 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-67d996989d-c5mcj" event={"ID":"dfcfffcb-36da-4853-8ed7-ba76caba9d3c","Type":"ContainerStarted","Data":"acd85a1f24d3b8bf6e07821ab0cfda3442b573cc2f992e69b425d0ce19425129"} Feb 27 16:40:39 crc kubenswrapper[4814]: I0227 16:40:39.920611 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-f4dv9" event={"ID":"71ebdc25-566e-4b42-8f48-6f3c52e1125c","Type":"ContainerStarted","Data":"019977d2ae32f8039520d04df02b4660d3e320656a563926032bec0ad9a5e9bb"} Feb 27 16:40:39 crc kubenswrapper[4814]: E0227 16:40:39.922200 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:71f2ab3bb41d1743287a3270dd49e32192b347d8ba7353d2250cbd7e8528219b\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-f4dv9" podUID="71ebdc25-566e-4b42-8f48-6f3c52e1125c" Feb 27 16:40:39 crc kubenswrapper[4814]: I0227 16:40:39.935043 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-hbdxk" event={"ID":"103a95f2-68f7-41b0-9dea-46cde1679842","Type":"ContainerStarted","Data":"e73139de008ef4fecc392e1d1e1bc9f8fa39a0f29e6d63fe21f25e6bc752a452"} Feb 27 16:40:39 crc kubenswrapper[4814]: I0227 16:40:39.976467 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-tck96" event={"ID":"30b75b23-8950-40e1-a515-eaf136867a25","Type":"ContainerStarted","Data":"1fee8ec8e70e77c7ff76a8ad1ad36d25c3fdcfe88410d6224f7cb7c69f86519f"} Feb 27 16:40:39 crc kubenswrapper[4814]: E0227 16:40:39.979172 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:9d03f03aa9a460f1fcac8875064808c03e4ecd0388873bbfb9c7dc58331f3968\\\"\"" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-tck96" podUID="30b75b23-8950-40e1-a515-eaf136867a25" Feb 27 16:40:39 crc kubenswrapper[4814]: I0227 16:40:39.990447 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-hh455" event={"ID":"4b02bae2-c5ac-423e-a2f1-b95de1264f47","Type":"ContainerStarted","Data":"ab0ee9c0aba5a7fcbb60238d6bb3d99de4aa8f9aa3aa840402510f146e0fe07a"} Feb 27 16:40:39 crc kubenswrapper[4814]: E0227 16:40:39.993616 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:06311600a491c689493552e7ff26e36df740fa4e7c143fca874bef19f24afb97\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-hh455" podUID="4b02bae2-c5ac-423e-a2f1-b95de1264f47" Feb 27 16:40:39 crc kubenswrapper[4814]: I0227 16:40:39.998728 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-5958h" event={"ID":"c6faf190-d834-45e6-8a18-067067c0bef5","Type":"ContainerStarted","Data":"0b0e284efa606aafd7137754bf1a9513e7e9830ef1aeb1c6525560e7c6c031db"} Feb 27 16:40:40 crc kubenswrapper[4814]: I0227 16:40:40.017428 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-ccrlj" event={"ID":"a3038d2f-e4de-4d0f-8ac0-e5c666077395","Type":"ContainerStarted","Data":"1b5172dd20c04b2a39744610d1315ba6f360e0ea34635b7f4f75a03cd59ba15c"} Feb 27 16:40:40 crc kubenswrapper[4814]: I0227 16:40:40.028436 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-4m2wv" event={"ID":"75ee2f25-f59a-4a60-8245-1b638957d234","Type":"ContainerStarted","Data":"fda72054dae41085ff85e692d2d4a04ece57bcf2f2ff707be94d460332bdee13"} Feb 27 16:40:40 crc kubenswrapper[4814]: I0227 16:40:40.504007 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afffcacd-d821-4cfd-aad8-462aafd2986f" path="/var/lib/kubelet/pods/afffcacd-d821-4cfd-aad8-462aafd2986f/volumes" Feb 27 16:40:40 crc kubenswrapper[4814]: I0227 16:40:40.684125 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-qc48n\" (UID: \"2d989d20-0a79-448b-8b73-ad9378119232\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" Feb 27 16:40:40 crc kubenswrapper[4814]: E0227 16:40:40.684296 4814 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 27 16:40:40 crc kubenswrapper[4814]: E0227 16:40:40.685475 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert podName:2d989d20-0a79-448b-8b73-ad9378119232 nodeName:}" failed. No retries permitted until 2026-02-27 16:40:44.685442302 +0000 UTC m=+1057.138067132 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert") pod "infra-operator-controller-manager-f7fcc58b9-qc48n" (UID: "2d989d20-0a79-448b-8b73-ad9378119232") : secret "infra-operator-webhook-server-cert" not found Feb 27 16:40:40 crc kubenswrapper[4814]: I0227 16:40:40.889721 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr\" (UID: \"edf51635-71a3-4c03-8aef-50212a3b8247\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" Feb 27 16:40:40 crc kubenswrapper[4814]: E0227 16:40:40.889955 4814 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 16:40:40 crc kubenswrapper[4814]: E0227 16:40:40.890095 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert podName:edf51635-71a3-4c03-8aef-50212a3b8247 nodeName:}" failed. No retries permitted until 2026-02-27 16:40:44.890050579 +0000 UTC m=+1057.342675409 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" (UID: "edf51635-71a3-4c03-8aef-50212a3b8247") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 16:40:41 crc kubenswrapper[4814]: E0227 16:40:41.076662 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:9d03f03aa9a460f1fcac8875064808c03e4ecd0388873bbfb9c7dc58331f3968\\\"\"" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-tck96" podUID="30b75b23-8950-40e1-a515-eaf136867a25" Feb 27 16:40:41 crc kubenswrapper[4814]: E0227 16:40:41.076918 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:1b9074a4ce16396d8bd2d30a475fc8c2f004f75a023e3eef8950661e89c0bcc6\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-4s2gh" podUID="3f85fa1d-a7c6-49b8-8f85-d02670fa2849" Feb 27 16:40:41 crc kubenswrapper[4814]: E0227 16:40:41.077308 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:114c0dee0bab1d453890e9dcc7727de749055bdbea049384d5696e7ac8d78fe3\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-p74f5" podUID="ea6ecef0-2796-4c32-9a9a-ee178574d2f9" Feb 27 16:40:41 crc kubenswrapper[4814]: E0227 16:40:41.079688 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:71f2ab3bb41d1743287a3270dd49e32192b347d8ba7353d2250cbd7e8528219b\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-f4dv9" podUID="71ebdc25-566e-4b42-8f48-6f3c52e1125c" Feb 27 16:40:41 crc kubenswrapper[4814]: E0227 16:40:41.079691 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:06311600a491c689493552e7ff26e36df740fa4e7c143fca874bef19f24afb97\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-hh455" podUID="4b02bae2-c5ac-423e-a2f1-b95de1264f47" Feb 27 16:40:41 crc kubenswrapper[4814]: E0227 16:40:41.080279 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwz52" podUID="60ad038c-aad7-4309-89cf-17bf1bda00ac" Feb 27 16:40:41 crc kubenswrapper[4814]: E0227 16:40:41.080986 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:f309cdea8084a4b1e8cbcd732d6e250fd93c55cfd1b48ba9026907c8591faab7\\\"\"" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-pq6mn" podUID="59cb5f17-969c-4c4f-81bf-a5b0e65acace" Feb 27 16:40:41 crc kubenswrapper[4814]: I0227 16:40:41.193535 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:41 crc kubenswrapper[4814]: E0227 16:40:41.193702 4814 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 16:40:41 crc kubenswrapper[4814]: E0227 16:40:41.193863 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs podName:96d3c178-c0f0-4bd5-b3bd-0e572404d30a nodeName:}" failed. No retries permitted until 2026-02-27 16:40:45.19384971 +0000 UTC m=+1057.646474540 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs") pod "openstack-operator-controller-manager-655dd5d898-d7f5g" (UID: "96d3c178-c0f0-4bd5-b3bd-0e572404d30a") : secret "webhook-server-cert" not found Feb 27 16:40:41 crc kubenswrapper[4814]: I0227 16:40:41.197181 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:41 crc kubenswrapper[4814]: E0227 16:40:41.197322 4814 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 16:40:41 crc kubenswrapper[4814]: E0227 16:40:41.197357 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs podName:96d3c178-c0f0-4bd5-b3bd-0e572404d30a nodeName:}" failed. No retries permitted until 2026-02-27 16:40:45.197346606 +0000 UTC m=+1057.649971436 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs") pod "openstack-operator-controller-manager-655dd5d898-d7f5g" (UID: "96d3c178-c0f0-4bd5-b3bd-0e572404d30a") : secret "metrics-server-cert" not found Feb 27 16:40:44 crc kubenswrapper[4814]: I0227 16:40:44.748063 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-qc48n\" (UID: \"2d989d20-0a79-448b-8b73-ad9378119232\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" Feb 27 16:40:44 crc kubenswrapper[4814]: E0227 16:40:44.748281 4814 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 27 16:40:44 crc kubenswrapper[4814]: E0227 16:40:44.748743 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert podName:2d989d20-0a79-448b-8b73-ad9378119232 nodeName:}" failed. No retries permitted until 2026-02-27 16:40:52.748708508 +0000 UTC m=+1065.201333338 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert") pod "infra-operator-controller-manager-f7fcc58b9-qc48n" (UID: "2d989d20-0a79-448b-8b73-ad9378119232") : secret "infra-operator-webhook-server-cert" not found Feb 27 16:40:44 crc kubenswrapper[4814]: I0227 16:40:44.952050 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr\" (UID: \"edf51635-71a3-4c03-8aef-50212a3b8247\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" Feb 27 16:40:44 crc kubenswrapper[4814]: E0227 16:40:44.952334 4814 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 16:40:44 crc kubenswrapper[4814]: E0227 16:40:44.952679 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert podName:edf51635-71a3-4c03-8aef-50212a3b8247 nodeName:}" failed. No retries permitted until 2026-02-27 16:40:52.952656966 +0000 UTC m=+1065.405281826 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" (UID: "edf51635-71a3-4c03-8aef-50212a3b8247") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 16:40:45 crc kubenswrapper[4814]: I0227 16:40:45.259381 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:45 crc kubenswrapper[4814]: I0227 16:40:45.259538 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:45 crc kubenswrapper[4814]: E0227 16:40:45.259550 4814 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 16:40:45 crc kubenswrapper[4814]: E0227 16:40:45.259611 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs podName:96d3c178-c0f0-4bd5-b3bd-0e572404d30a nodeName:}" failed. No retries permitted until 2026-02-27 16:40:53.259595982 +0000 UTC m=+1065.712220812 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs") pod "openstack-operator-controller-manager-655dd5d898-d7f5g" (UID: "96d3c178-c0f0-4bd5-b3bd-0e572404d30a") : secret "metrics-server-cert" not found Feb 27 16:40:45 crc kubenswrapper[4814]: E0227 16:40:45.259863 4814 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 16:40:45 crc kubenswrapper[4814]: E0227 16:40:45.259975 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs podName:96d3c178-c0f0-4bd5-b3bd-0e572404d30a nodeName:}" failed. No retries permitted until 2026-02-27 16:40:53.259953753 +0000 UTC m=+1065.712578593 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs") pod "openstack-operator-controller-manager-655dd5d898-d7f5g" (UID: "96d3c178-c0f0-4bd5-b3bd-0e572404d30a") : secret "webhook-server-cert" not found Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.145458 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-nxrtn" event={"ID":"9b633770-3463-40ee-b5fc-48ec5fb4846c","Type":"ContainerStarted","Data":"dafb17554295011006520e1d8eab1023df715d1d84f00a99b47c839a7693e39d"} Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.145964 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-nxrtn" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.147906 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-m5zqv" event={"ID":"5ee5efe1-368e-476e-8516-d9b81d9a38a6","Type":"ContainerStarted","Data":"795c314f7c13a712cf8282e0c64fc99f67de216ea78cff9c4534e57f59cfec25"} Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.148456 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-m5zqv" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.149399 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-pjncz" event={"ID":"6ee0f0c6-52ee-4d20-999f-a0a579dcdaef","Type":"ContainerStarted","Data":"e8bad409de4f3b8e0adb358d02bafad5d1b2a7d9365f38eb810c2046dfe58014"} Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.149728 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-pjncz" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.150627 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54688575f-q2gkq" event={"ID":"50d18356-e250-46b6-bcd3-053f770f8f58","Type":"ContainerStarted","Data":"f445ca4e49826547577631d1526c1c7e83120646e9fa7cf3cf895d4e4e0a173a"} Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.150964 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-54688575f-q2gkq" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.152035 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-9kqgh" event={"ID":"49b03138-9b4c-486c-af91-37c16e8c2536","Type":"ContainerStarted","Data":"30ff23c9583e6bcbf61ddaf84291d5306459cd97e6d9f67cb4a59cbb935cd4fb"} Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.152396 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-9kqgh" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.153292 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-4m2wv" event={"ID":"75ee2f25-f59a-4a60-8245-1b638957d234","Type":"ContainerStarted","Data":"e216c7e0a69e1f8515b7f2b23a5b06bca70902c149d3a77a285f00d511e92c89"} Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.153603 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-4m2wv" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.154430 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-67d996989d-c5mcj" event={"ID":"dfcfffcb-36da-4853-8ed7-ba76caba9d3c","Type":"ContainerStarted","Data":"7465eaa09119477bc59791da84a9d315856c1e5b672de1ceeae874a0d8313e8f"} Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.154761 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-67d996989d-c5mcj" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.155601 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-hbdxk" event={"ID":"103a95f2-68f7-41b0-9dea-46cde1679842","Type":"ContainerStarted","Data":"13fdfa0c51d7be4a8167eed7463c2b148422426982bbe58aed8ebe08c4ba04d1"} Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.155923 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-hbdxk" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.156786 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-fp474" event={"ID":"2192fb73-bd5d-440f-8c01-2871658164d9","Type":"ContainerStarted","Data":"f1fc65cc93e681b4af7dbfd4c96ff888e61bfd731439864d6373369c74c67cb5"} Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.157100 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-fp474" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.157973 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-xnhhg" event={"ID":"6ede514a-7a77-4cfe-a746-3d7bd2f5d7fd","Type":"ContainerStarted","Data":"5ec2a551554bda60cfb0cbe8589e6bea7f4b73762c369f153474ae35911520c3"} Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.158305 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-xnhhg" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.159163 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-5958h" event={"ID":"c6faf190-d834-45e6-8a18-067067c0bef5","Type":"ContainerStarted","Data":"12db55cdd16bce08f522be1adb1c41805622a7e73d0c5f5976e419bc475bc0a9"} Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.159414 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-5958h" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.160594 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-2d2n7" event={"ID":"de6e87db-4283-4aab-a97e-23792fd3b480","Type":"ContainerStarted","Data":"e91ddd0b71de04a832fb46c5e6c0c24b03024cdfa08a915562c403d0a881ae4a"} Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.160705 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-2d2n7" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.162565 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-ccrlj" event={"ID":"a3038d2f-e4de-4d0f-8ac0-e5c666077395","Type":"ContainerStarted","Data":"7b53c6820eb4fd4203b32c83bb093ec3005250d7bc35a165df342f74b2ee1830"} Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.162703 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-ccrlj" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.235091 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-nxrtn" podStartSLOduration=3.890485247 podStartE2EDuration="15.235075613s" podCreationTimestamp="2026-02-27 16:40:36 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.796445373 +0000 UTC m=+1051.249070203" lastFinishedPulling="2026-02-27 16:40:50.141035749 +0000 UTC m=+1062.593660569" observedRunningTime="2026-02-27 16:40:51.234440314 +0000 UTC m=+1063.687065144" watchObservedRunningTime="2026-02-27 16:40:51.235075613 +0000 UTC m=+1063.687700443" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.305841 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-ccrlj" podStartSLOduration=4.001099146 podStartE2EDuration="15.305822402s" podCreationTimestamp="2026-02-27 16:40:36 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.800498755 +0000 UTC m=+1051.253123585" lastFinishedPulling="2026-02-27 16:40:50.105222011 +0000 UTC m=+1062.557846841" observedRunningTime="2026-02-27 16:40:51.304947986 +0000 UTC m=+1063.757572817" watchObservedRunningTime="2026-02-27 16:40:51.305822402 +0000 UTC m=+1063.758447232" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.347923 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-4m2wv" podStartSLOduration=4.034355637 podStartE2EDuration="15.347905561s" podCreationTimestamp="2026-02-27 16:40:36 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.798527146 +0000 UTC m=+1051.251151976" lastFinishedPulling="2026-02-27 16:40:50.11207703 +0000 UTC m=+1062.564701900" observedRunningTime="2026-02-27 16:40:51.345188649 +0000 UTC m=+1063.797813479" watchObservedRunningTime="2026-02-27 16:40:51.347905561 +0000 UTC m=+1063.800530391" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.410734 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-hbdxk" podStartSLOduration=3.080419928 podStartE2EDuration="14.41071464s" podCreationTimestamp="2026-02-27 16:40:37 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.779676673 +0000 UTC m=+1051.232301503" lastFinishedPulling="2026-02-27 16:40:50.109971385 +0000 UTC m=+1062.562596215" observedRunningTime="2026-02-27 16:40:51.40840385 +0000 UTC m=+1063.861028680" watchObservedRunningTime="2026-02-27 16:40:51.41071464 +0000 UTC m=+1063.863339470" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.439992 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-pjncz" podStartSLOduration=3.7154075559999997 podStartE2EDuration="15.439977639s" podCreationTimestamp="2026-02-27 16:40:36 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.322858422 +0000 UTC m=+1050.775483252" lastFinishedPulling="2026-02-27 16:40:50.047428505 +0000 UTC m=+1062.500053335" observedRunningTime="2026-02-27 16:40:51.438523405 +0000 UTC m=+1063.891148235" watchObservedRunningTime="2026-02-27 16:40:51.439977639 +0000 UTC m=+1063.892602469" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.462223 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-9kqgh" podStartSLOduration=3.673682528 podStartE2EDuration="15.462208234s" podCreationTimestamp="2026-02-27 16:40:36 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.340315353 +0000 UTC m=+1050.792940183" lastFinishedPulling="2026-02-27 16:40:50.128841049 +0000 UTC m=+1062.581465889" observedRunningTime="2026-02-27 16:40:51.459107771 +0000 UTC m=+1063.911732601" watchObservedRunningTime="2026-02-27 16:40:51.462208234 +0000 UTC m=+1063.914833064" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.506971 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-m5zqv" podStartSLOduration=4.05616745 podStartE2EDuration="15.506950144s" podCreationTimestamp="2026-02-27 16:40:36 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.776229668 +0000 UTC m=+1051.228854498" lastFinishedPulling="2026-02-27 16:40:50.227012362 +0000 UTC m=+1062.679637192" observedRunningTime="2026-02-27 16:40:51.50254022 +0000 UTC m=+1063.955165070" watchObservedRunningTime="2026-02-27 16:40:51.506950144 +0000 UTC m=+1063.959574974" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.549964 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-2d2n7" podStartSLOduration=3.2329810979999998 podStartE2EDuration="15.549944631s" podCreationTimestamp="2026-02-27 16:40:36 +0000 UTC" firstStartedPulling="2026-02-27 16:40:37.793407254 +0000 UTC m=+1050.246032084" lastFinishedPulling="2026-02-27 16:40:50.110370747 +0000 UTC m=+1062.562995617" observedRunningTime="2026-02-27 16:40:51.546575898 +0000 UTC m=+1063.999200728" watchObservedRunningTime="2026-02-27 16:40:51.549944631 +0000 UTC m=+1064.002569461" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.574560 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-67d996989d-c5mcj" podStartSLOduration=4.305927469 podStartE2EDuration="15.574540968s" podCreationTimestamp="2026-02-27 16:40:36 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.816361567 +0000 UTC m=+1051.268986397" lastFinishedPulling="2026-02-27 16:40:50.084975026 +0000 UTC m=+1062.537599896" observedRunningTime="2026-02-27 16:40:51.572103714 +0000 UTC m=+1064.024728564" watchObservedRunningTime="2026-02-27 16:40:51.574540968 +0000 UTC m=+1064.027165798" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.604018 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-fp474" podStartSLOduration=4.317600565 podStartE2EDuration="15.603998493s" podCreationTimestamp="2026-02-27 16:40:36 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.779661873 +0000 UTC m=+1051.232286703" lastFinishedPulling="2026-02-27 16:40:50.066059791 +0000 UTC m=+1062.518684631" observedRunningTime="2026-02-27 16:40:51.600205198 +0000 UTC m=+1064.052830028" watchObservedRunningTime="2026-02-27 16:40:51.603998493 +0000 UTC m=+1064.056623323" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.618638 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-5958h" podStartSLOduration=4.357316551 podStartE2EDuration="15.618616498s" podCreationTimestamp="2026-02-27 16:40:36 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.78123667 +0000 UTC m=+1051.233861500" lastFinishedPulling="2026-02-27 16:40:50.042536617 +0000 UTC m=+1062.495161447" observedRunningTime="2026-02-27 16:40:51.615521404 +0000 UTC m=+1064.068146234" watchObservedRunningTime="2026-02-27 16:40:51.618616498 +0000 UTC m=+1064.071241328" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.648565 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-54688575f-q2gkq" podStartSLOduration=4.350765762 podStartE2EDuration="15.648543147s" podCreationTimestamp="2026-02-27 16:40:36 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.772129203 +0000 UTC m=+1051.224754033" lastFinishedPulling="2026-02-27 16:40:50.069906588 +0000 UTC m=+1062.522531418" observedRunningTime="2026-02-27 16:40:51.64439994 +0000 UTC m=+1064.097024770" watchObservedRunningTime="2026-02-27 16:40:51.648543147 +0000 UTC m=+1064.101167977" Feb 27 16:40:51 crc kubenswrapper[4814]: I0227 16:40:51.664705 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-xnhhg" podStartSLOduration=3.952261013 podStartE2EDuration="15.664693687s" podCreationTimestamp="2026-02-27 16:40:36 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.398869032 +0000 UTC m=+1050.851493862" lastFinishedPulling="2026-02-27 16:40:50.111301706 +0000 UTC m=+1062.563926536" observedRunningTime="2026-02-27 16:40:51.662373607 +0000 UTC m=+1064.114998437" watchObservedRunningTime="2026-02-27 16:40:51.664693687 +0000 UTC m=+1064.117318517" Feb 27 16:40:52 crc kubenswrapper[4814]: I0227 16:40:52.770184 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-qc48n\" (UID: \"2d989d20-0a79-448b-8b73-ad9378119232\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" Feb 27 16:40:52 crc kubenswrapper[4814]: I0227 16:40:52.778823 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d989d20-0a79-448b-8b73-ad9378119232-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-qc48n\" (UID: \"2d989d20-0a79-448b-8b73-ad9378119232\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" Feb 27 16:40:52 crc kubenswrapper[4814]: I0227 16:40:52.972931 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr\" (UID: \"edf51635-71a3-4c03-8aef-50212a3b8247\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" Feb 27 16:40:52 crc kubenswrapper[4814]: I0227 16:40:52.977983 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/edf51635-71a3-4c03-8aef-50212a3b8247-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr\" (UID: \"edf51635-71a3-4c03-8aef-50212a3b8247\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" Feb 27 16:40:53 crc kubenswrapper[4814]: I0227 16:40:53.001715 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" Feb 27 16:40:53 crc kubenswrapper[4814]: I0227 16:40:53.207446 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" Feb 27 16:40:53 crc kubenswrapper[4814]: I0227 16:40:53.280020 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:53 crc kubenswrapper[4814]: I0227 16:40:53.280213 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:53 crc kubenswrapper[4814]: E0227 16:40:53.284446 4814 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 16:40:53 crc kubenswrapper[4814]: E0227 16:40:53.284489 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs podName:96d3c178-c0f0-4bd5-b3bd-0e572404d30a nodeName:}" failed. No retries permitted until 2026-02-27 16:41:09.284475607 +0000 UTC m=+1081.737100437 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs") pod "openstack-operator-controller-manager-655dd5d898-d7f5g" (UID: "96d3c178-c0f0-4bd5-b3bd-0e572404d30a") : secret "webhook-server-cert" not found Feb 27 16:40:53 crc kubenswrapper[4814]: I0227 16:40:53.288924 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-metrics-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:40:53 crc kubenswrapper[4814]: I0227 16:40:53.514473 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n"] Feb 27 16:40:53 crc kubenswrapper[4814]: W0227 16:40:53.520195 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d989d20_0a79_448b_8b73_ad9378119232.slice/crio-54cf1e0bb974bb3b6870f0e8bf8fb0589c8ef51e4fade8f2b9665e1a0973f313 WatchSource:0}: Error finding container 54cf1e0bb974bb3b6870f0e8bf8fb0589c8ef51e4fade8f2b9665e1a0973f313: Status 404 returned error can't find the container with id 54cf1e0bb974bb3b6870f0e8bf8fb0589c8ef51e4fade8f2b9665e1a0973f313 Feb 27 16:40:53 crc kubenswrapper[4814]: I0227 16:40:53.659008 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr"] Feb 27 16:40:53 crc kubenswrapper[4814]: W0227 16:40:53.667003 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedf51635_71a3_4c03_8aef_50212a3b8247.slice/crio-10112880fc7c67183b40e59f3e67485509ea104659b157c0f79fd79d1ba50940 WatchSource:0}: Error finding container 10112880fc7c67183b40e59f3e67485509ea104659b157c0f79fd79d1ba50940: Status 404 returned error can't find the container with id 10112880fc7c67183b40e59f3e67485509ea104659b157c0f79fd79d1ba50940 Feb 27 16:40:54 crc kubenswrapper[4814]: I0227 16:40:54.200169 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" event={"ID":"2d989d20-0a79-448b-8b73-ad9378119232","Type":"ContainerStarted","Data":"54cf1e0bb974bb3b6870f0e8bf8fb0589c8ef51e4fade8f2b9665e1a0973f313"} Feb 27 16:40:54 crc kubenswrapper[4814]: I0227 16:40:54.201616 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" event={"ID":"edf51635-71a3-4c03-8aef-50212a3b8247","Type":"ContainerStarted","Data":"10112880fc7c67183b40e59f3e67485509ea104659b157c0f79fd79d1ba50940"} Feb 27 16:40:57 crc kubenswrapper[4814]: I0227 16:40:57.059976 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-xnhhg" Feb 27 16:40:57 crc kubenswrapper[4814]: I0227 16:40:57.081218 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-2d2n7" Feb 27 16:40:57 crc kubenswrapper[4814]: I0227 16:40:57.106537 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-pjncz" Feb 27 16:40:57 crc kubenswrapper[4814]: I0227 16:40:57.131670 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-9kqgh" Feb 27 16:40:57 crc kubenswrapper[4814]: I0227 16:40:57.184884 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-ccrlj" Feb 27 16:40:57 crc kubenswrapper[4814]: I0227 16:40:57.268386 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-54688575f-q2gkq" Feb 27 16:40:57 crc kubenswrapper[4814]: I0227 16:40:57.280074 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-m5zqv" Feb 27 16:40:57 crc kubenswrapper[4814]: I0227 16:40:57.296875 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-4m2wv" Feb 27 16:40:57 crc kubenswrapper[4814]: I0227 16:40:57.390860 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-5958h" Feb 27 16:40:57 crc kubenswrapper[4814]: I0227 16:40:57.396399 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-hbdxk" Feb 27 16:40:57 crc kubenswrapper[4814]: I0227 16:40:57.433054 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-fp474" Feb 27 16:40:57 crc kubenswrapper[4814]: I0227 16:40:57.435692 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-nxrtn" Feb 27 16:40:57 crc kubenswrapper[4814]: I0227 16:40:57.485657 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-67d996989d-c5mcj" Feb 27 16:41:06 crc kubenswrapper[4814]: I0227 16:41:06.309109 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-hh455" event={"ID":"4b02bae2-c5ac-423e-a2f1-b95de1264f47","Type":"ContainerStarted","Data":"b86057da7f7f41f62a2d4f402227652f2c68bc0b6b9f61aa7ce3634803ff1e32"} Feb 27 16:41:06 crc kubenswrapper[4814]: I0227 16:41:06.309843 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-hh455" Feb 27 16:41:06 crc kubenswrapper[4814]: I0227 16:41:06.326622 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwz52" event={"ID":"60ad038c-aad7-4309-89cf-17bf1bda00ac","Type":"ContainerStarted","Data":"74bdd12a5cd273e0e6d1e81c5d790c30daeabf6bc6f84101522f57417337035e"} Feb 27 16:41:06 crc kubenswrapper[4814]: I0227 16:41:06.333134 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-hh455" podStartSLOduration=2.257603665 podStartE2EDuration="29.333118723s" podCreationTimestamp="2026-02-27 16:40:37 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.882059964 +0000 UTC m=+1051.334684794" lastFinishedPulling="2026-02-27 16:41:05.957575002 +0000 UTC m=+1078.410199852" observedRunningTime="2026-02-27 16:41:06.332936408 +0000 UTC m=+1078.785561238" watchObservedRunningTime="2026-02-27 16:41:06.333118723 +0000 UTC m=+1078.785743553" Feb 27 16:41:06 crc kubenswrapper[4814]: I0227 16:41:06.340336 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-pq6mn" event={"ID":"59cb5f17-969c-4c4f-81bf-a5b0e65acace","Type":"ContainerStarted","Data":"f49ac72bc97ab69ef218c6724adefb83b83d2f4781d7a888454ef63ddfc6659e"} Feb 27 16:41:06 crc kubenswrapper[4814]: I0227 16:41:06.340686 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-pq6mn" Feb 27 16:41:06 crc kubenswrapper[4814]: I0227 16:41:06.343945 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-4s2gh" event={"ID":"3f85fa1d-a7c6-49b8-8f85-d02670fa2849","Type":"ContainerStarted","Data":"a99fb918c76c5f688aa6ee363ae21e756a7cfefa1ae5fb945f42d2116b7c6366"} Feb 27 16:41:06 crc kubenswrapper[4814]: I0227 16:41:06.344159 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-4s2gh" Feb 27 16:41:06 crc kubenswrapper[4814]: I0227 16:41:06.354130 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwz52" podStartSLOduration=2.202419039 podStartE2EDuration="29.354115641s" podCreationTimestamp="2026-02-27 16:40:37 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.845697309 +0000 UTC m=+1051.298322139" lastFinishedPulling="2026-02-27 16:41:05.997393871 +0000 UTC m=+1078.450018741" observedRunningTime="2026-02-27 16:41:06.352383628 +0000 UTC m=+1078.805008458" watchObservedRunningTime="2026-02-27 16:41:06.354115641 +0000 UTC m=+1078.806740471" Feb 27 16:41:06 crc kubenswrapper[4814]: I0227 16:41:06.366640 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-pq6mn" podStartSLOduration=2.265100763 podStartE2EDuration="29.366627151s" podCreationTimestamp="2026-02-27 16:40:37 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.857617361 +0000 UTC m=+1051.310242191" lastFinishedPulling="2026-02-27 16:41:05.959143739 +0000 UTC m=+1078.411768579" observedRunningTime="2026-02-27 16:41:06.365749294 +0000 UTC m=+1078.818374144" watchObservedRunningTime="2026-02-27 16:41:06.366627151 +0000 UTC m=+1078.819251981" Feb 27 16:41:06 crc kubenswrapper[4814]: I0227 16:41:06.381066 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-4s2gh" podStartSLOduration=2.783623219 podStartE2EDuration="29.381048909s" podCreationTimestamp="2026-02-27 16:40:37 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.845790192 +0000 UTC m=+1051.298415022" lastFinishedPulling="2026-02-27 16:41:05.443215842 +0000 UTC m=+1077.895840712" observedRunningTime="2026-02-27 16:41:06.379961716 +0000 UTC m=+1078.832586546" watchObservedRunningTime="2026-02-27 16:41:06.381048909 +0000 UTC m=+1078.833673759" Feb 27 16:41:07 crc kubenswrapper[4814]: I0227 16:41:07.349745 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" event={"ID":"2d989d20-0a79-448b-8b73-ad9378119232","Type":"ContainerStarted","Data":"f44e6b6a3b5b05fa07338763a5a81e0377f924fbc800c50ab3414a47072955e9"} Feb 27 16:41:07 crc kubenswrapper[4814]: I0227 16:41:07.350508 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" Feb 27 16:41:07 crc kubenswrapper[4814]: I0227 16:41:07.351790 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-p74f5" event={"ID":"ea6ecef0-2796-4c32-9a9a-ee178574d2f9","Type":"ContainerStarted","Data":"1701e8c3f99590e143d05b32010beaa4abf189d194d79235c80ec1ed93917e5b"} Feb 27 16:41:07 crc kubenswrapper[4814]: I0227 16:41:07.352291 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-p74f5" Feb 27 16:41:07 crc kubenswrapper[4814]: I0227 16:41:07.353815 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-f4dv9" event={"ID":"71ebdc25-566e-4b42-8f48-6f3c52e1125c","Type":"ContainerStarted","Data":"3e60f6c556d45da2780d27dd8976a8fe3daa6669d60deeb2840e92eee61158f1"} Feb 27 16:41:07 crc kubenswrapper[4814]: I0227 16:41:07.354000 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-f4dv9" Feb 27 16:41:07 crc kubenswrapper[4814]: I0227 16:41:07.355204 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-tck96" event={"ID":"30b75b23-8950-40e1-a515-eaf136867a25","Type":"ContainerStarted","Data":"43639a8fc135133d4c5ea261a762805d9ea0a600b4b517fbaedc6aad9bc173a2"} Feb 27 16:41:07 crc kubenswrapper[4814]: I0227 16:41:07.355501 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-tck96" Feb 27 16:41:07 crc kubenswrapper[4814]: I0227 16:41:07.356698 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" event={"ID":"edf51635-71a3-4c03-8aef-50212a3b8247","Type":"ContainerStarted","Data":"3effc13a9a1326145c4d3f9f4c7f8365b677d54e52e256119252e0b94f36ec0b"} Feb 27 16:41:07 crc kubenswrapper[4814]: I0227 16:41:07.357362 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" Feb 27 16:41:07 crc kubenswrapper[4814]: I0227 16:41:07.369095 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" podStartSLOduration=19.449415871 podStartE2EDuration="31.369068102s" podCreationTimestamp="2026-02-27 16:40:36 +0000 UTC" firstStartedPulling="2026-02-27 16:40:53.523470938 +0000 UTC m=+1065.976095778" lastFinishedPulling="2026-02-27 16:41:05.443123179 +0000 UTC m=+1077.895748009" observedRunningTime="2026-02-27 16:41:07.368773063 +0000 UTC m=+1079.821397903" watchObservedRunningTime="2026-02-27 16:41:07.369068102 +0000 UTC m=+1079.821692952" Feb 27 16:41:07 crc kubenswrapper[4814]: I0227 16:41:07.389927 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-p74f5" podStartSLOduration=4.237143228 podStartE2EDuration="31.389912304s" podCreationTimestamp="2026-02-27 16:40:36 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.816428429 +0000 UTC m=+1051.269053259" lastFinishedPulling="2026-02-27 16:41:05.969197465 +0000 UTC m=+1078.421822335" observedRunningTime="2026-02-27 16:41:07.386994256 +0000 UTC m=+1079.839619076" watchObservedRunningTime="2026-02-27 16:41:07.389912304 +0000 UTC m=+1079.842537134" Feb 27 16:41:07 crc kubenswrapper[4814]: I0227 16:41:07.415718 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-f4dv9" podStartSLOduration=4.314469119 podStartE2EDuration="31.415700908s" podCreationTimestamp="2026-02-27 16:40:36 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.857414085 +0000 UTC m=+1051.310038915" lastFinishedPulling="2026-02-27 16:41:05.958645854 +0000 UTC m=+1078.411270704" observedRunningTime="2026-02-27 16:41:07.408930962 +0000 UTC m=+1079.861555802" watchObservedRunningTime="2026-02-27 16:41:07.415700908 +0000 UTC m=+1079.868325728" Feb 27 16:41:07 crc kubenswrapper[4814]: I0227 16:41:07.439535 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" podStartSLOduration=19.176648232 podStartE2EDuration="31.439516591s" podCreationTimestamp="2026-02-27 16:40:36 +0000 UTC" firstStartedPulling="2026-02-27 16:40:53.670749754 +0000 UTC m=+1066.123374594" lastFinishedPulling="2026-02-27 16:41:05.933618103 +0000 UTC m=+1078.386242953" observedRunningTime="2026-02-27 16:41:07.434549151 +0000 UTC m=+1079.887173991" watchObservedRunningTime="2026-02-27 16:41:07.439516591 +0000 UTC m=+1079.892141421" Feb 27 16:41:07 crc kubenswrapper[4814]: I0227 16:41:07.459338 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-tck96" podStartSLOduration=4.769264024 podStartE2EDuration="30.459319533s" podCreationTimestamp="2026-02-27 16:40:37 +0000 UTC" firstStartedPulling="2026-02-27 16:40:38.845866064 +0000 UTC m=+1051.298490894" lastFinishedPulling="2026-02-27 16:41:04.535921573 +0000 UTC m=+1076.988546403" observedRunningTime="2026-02-27 16:41:07.45229577 +0000 UTC m=+1079.904920600" watchObservedRunningTime="2026-02-27 16:41:07.459319533 +0000 UTC m=+1079.911944373" Feb 27 16:41:09 crc kubenswrapper[4814]: I0227 16:41:09.364577 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:41:09 crc kubenswrapper[4814]: I0227 16:41:09.373991 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/96d3c178-c0f0-4bd5-b3bd-0e572404d30a-webhook-certs\") pod \"openstack-operator-controller-manager-655dd5d898-d7f5g\" (UID: \"96d3c178-c0f0-4bd5-b3bd-0e572404d30a\") " pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:41:09 crc kubenswrapper[4814]: I0227 16:41:09.517560 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-xjqww" Feb 27 16:41:09 crc kubenswrapper[4814]: I0227 16:41:09.525512 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:41:10 crc kubenswrapper[4814]: I0227 16:41:10.041848 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g"] Feb 27 16:41:10 crc kubenswrapper[4814]: I0227 16:41:10.393361 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" event={"ID":"96d3c178-c0f0-4bd5-b3bd-0e572404d30a","Type":"ContainerStarted","Data":"78e4c73cd68de7ba15e73b49df75f6d5ebd35e5ae30cfddb14859516204bb58d"} Feb 27 16:41:10 crc kubenswrapper[4814]: I0227 16:41:10.393451 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" event={"ID":"96d3c178-c0f0-4bd5-b3bd-0e572404d30a","Type":"ContainerStarted","Data":"e0891c40191e06d3ee5bc2a57f709d7b135cb8881e246c7adc691184425c2861"} Feb 27 16:41:10 crc kubenswrapper[4814]: I0227 16:41:10.394788 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:41:13 crc kubenswrapper[4814]: I0227 16:41:13.019211 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-qc48n" Feb 27 16:41:13 crc kubenswrapper[4814]: I0227 16:41:13.064182 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" podStartSLOduration=36.064157593 podStartE2EDuration="36.064157593s" podCreationTimestamp="2026-02-27 16:40:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:41:10.44315298 +0000 UTC m=+1082.895777850" watchObservedRunningTime="2026-02-27 16:41:13.064157593 +0000 UTC m=+1085.516782453" Feb 27 16:41:13 crc kubenswrapper[4814]: I0227 16:41:13.220946 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr" Feb 27 16:41:17 crc kubenswrapper[4814]: I0227 16:41:17.202060 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-p74f5" Feb 27 16:41:17 crc kubenswrapper[4814]: I0227 16:41:17.478596 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-pq6mn" Feb 27 16:41:17 crc kubenswrapper[4814]: I0227 16:41:17.504750 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-4s2gh" Feb 27 16:41:17 crc kubenswrapper[4814]: I0227 16:41:17.523619 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-tck96" Feb 27 16:41:17 crc kubenswrapper[4814]: I0227 16:41:17.530078 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-f4dv9" Feb 27 16:41:18 crc kubenswrapper[4814]: I0227 16:41:18.092777 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-hh455" Feb 27 16:41:19 crc kubenswrapper[4814]: I0227 16:41:19.536461 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-655dd5d898-d7f5g" Feb 27 16:41:22 crc kubenswrapper[4814]: I0227 16:41:22.903043 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:41:22 crc kubenswrapper[4814]: I0227 16:41:22.903611 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.722685 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-k7ct4"] Feb 27 16:41:41 crc kubenswrapper[4814]: E0227 16:41:41.723582 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afffcacd-d821-4cfd-aad8-462aafd2986f" containerName="extract-content" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.723601 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="afffcacd-d821-4cfd-aad8-462aafd2986f" containerName="extract-content" Feb 27 16:41:41 crc kubenswrapper[4814]: E0227 16:41:41.723634 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afffcacd-d821-4cfd-aad8-462aafd2986f" containerName="extract-utilities" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.723644 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="afffcacd-d821-4cfd-aad8-462aafd2986f" containerName="extract-utilities" Feb 27 16:41:41 crc kubenswrapper[4814]: E0227 16:41:41.723659 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afffcacd-d821-4cfd-aad8-462aafd2986f" containerName="registry-server" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.723667 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="afffcacd-d821-4cfd-aad8-462aafd2986f" containerName="registry-server" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.723848 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="afffcacd-d821-4cfd-aad8-462aafd2986f" containerName="registry-server" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.724692 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-k7ct4" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.726436 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-47s6f" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.727151 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.727320 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.727467 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.730644 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-k7ct4"] Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.764644 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h7zs\" (UniqueName: \"kubernetes.io/projected/6faabeb6-6038-4bbc-8aae-3713e26566d2-kube-api-access-9h7zs\") pod \"dnsmasq-dns-675f4bcbfc-k7ct4\" (UID: \"6faabeb6-6038-4bbc-8aae-3713e26566d2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-k7ct4" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.764696 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6faabeb6-6038-4bbc-8aae-3713e26566d2-config\") pod \"dnsmasq-dns-675f4bcbfc-k7ct4\" (UID: \"6faabeb6-6038-4bbc-8aae-3713e26566d2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-k7ct4" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.788521 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-7c7nq"] Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.789604 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-7c7nq" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.792107 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.803583 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-7c7nq"] Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.865703 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t84hk\" (UniqueName: \"kubernetes.io/projected/2183292c-2d97-4a2c-bc26-a46a0ffbf019-kube-api-access-t84hk\") pod \"dnsmasq-dns-78dd6ddcc-7c7nq\" (UID: \"2183292c-2d97-4a2c-bc26-a46a0ffbf019\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7c7nq" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.865752 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h7zs\" (UniqueName: \"kubernetes.io/projected/6faabeb6-6038-4bbc-8aae-3713e26566d2-kube-api-access-9h7zs\") pod \"dnsmasq-dns-675f4bcbfc-k7ct4\" (UID: \"6faabeb6-6038-4bbc-8aae-3713e26566d2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-k7ct4" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.865835 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6faabeb6-6038-4bbc-8aae-3713e26566d2-config\") pod \"dnsmasq-dns-675f4bcbfc-k7ct4\" (UID: \"6faabeb6-6038-4bbc-8aae-3713e26566d2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-k7ct4" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.865921 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2183292c-2d97-4a2c-bc26-a46a0ffbf019-config\") pod \"dnsmasq-dns-78dd6ddcc-7c7nq\" (UID: \"2183292c-2d97-4a2c-bc26-a46a0ffbf019\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7c7nq" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.865981 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2183292c-2d97-4a2c-bc26-a46a0ffbf019-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-7c7nq\" (UID: \"2183292c-2d97-4a2c-bc26-a46a0ffbf019\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7c7nq" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.866839 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6faabeb6-6038-4bbc-8aae-3713e26566d2-config\") pod \"dnsmasq-dns-675f4bcbfc-k7ct4\" (UID: \"6faabeb6-6038-4bbc-8aae-3713e26566d2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-k7ct4" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.888015 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h7zs\" (UniqueName: \"kubernetes.io/projected/6faabeb6-6038-4bbc-8aae-3713e26566d2-kube-api-access-9h7zs\") pod \"dnsmasq-dns-675f4bcbfc-k7ct4\" (UID: \"6faabeb6-6038-4bbc-8aae-3713e26566d2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-k7ct4" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.967174 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2183292c-2d97-4a2c-bc26-a46a0ffbf019-config\") pod \"dnsmasq-dns-78dd6ddcc-7c7nq\" (UID: \"2183292c-2d97-4a2c-bc26-a46a0ffbf019\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7c7nq" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.967780 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2183292c-2d97-4a2c-bc26-a46a0ffbf019-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-7c7nq\" (UID: \"2183292c-2d97-4a2c-bc26-a46a0ffbf019\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7c7nq" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.967905 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t84hk\" (UniqueName: \"kubernetes.io/projected/2183292c-2d97-4a2c-bc26-a46a0ffbf019-kube-api-access-t84hk\") pod \"dnsmasq-dns-78dd6ddcc-7c7nq\" (UID: \"2183292c-2d97-4a2c-bc26-a46a0ffbf019\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7c7nq" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.967976 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2183292c-2d97-4a2c-bc26-a46a0ffbf019-config\") pod \"dnsmasq-dns-78dd6ddcc-7c7nq\" (UID: \"2183292c-2d97-4a2c-bc26-a46a0ffbf019\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7c7nq" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.969023 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2183292c-2d97-4a2c-bc26-a46a0ffbf019-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-7c7nq\" (UID: \"2183292c-2d97-4a2c-bc26-a46a0ffbf019\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7c7nq" Feb 27 16:41:41 crc kubenswrapper[4814]: I0227 16:41:41.985300 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t84hk\" (UniqueName: \"kubernetes.io/projected/2183292c-2d97-4a2c-bc26-a46a0ffbf019-kube-api-access-t84hk\") pod \"dnsmasq-dns-78dd6ddcc-7c7nq\" (UID: \"2183292c-2d97-4a2c-bc26-a46a0ffbf019\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7c7nq" Feb 27 16:41:42 crc kubenswrapper[4814]: I0227 16:41:42.040528 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-k7ct4" Feb 27 16:41:42 crc kubenswrapper[4814]: I0227 16:41:42.107003 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-7c7nq" Feb 27 16:41:42 crc kubenswrapper[4814]: I0227 16:41:42.518200 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-k7ct4"] Feb 27 16:41:42 crc kubenswrapper[4814]: I0227 16:41:42.595886 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-7c7nq"] Feb 27 16:41:42 crc kubenswrapper[4814]: W0227 16:41:42.601439 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2183292c_2d97_4a2c_bc26_a46a0ffbf019.slice/crio-2a39b6fc3f3d8ff628b145c26562265d936ea6d5910b804d4c61514f85e99675 WatchSource:0}: Error finding container 2a39b6fc3f3d8ff628b145c26562265d936ea6d5910b804d4c61514f85e99675: Status 404 returned error can't find the container with id 2a39b6fc3f3d8ff628b145c26562265d936ea6d5910b804d4c61514f85e99675 Feb 27 16:41:42 crc kubenswrapper[4814]: I0227 16:41:42.692739 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-7c7nq" event={"ID":"2183292c-2d97-4a2c-bc26-a46a0ffbf019","Type":"ContainerStarted","Data":"2a39b6fc3f3d8ff628b145c26562265d936ea6d5910b804d4c61514f85e99675"} Feb 27 16:41:42 crc kubenswrapper[4814]: I0227 16:41:42.694839 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-k7ct4" event={"ID":"6faabeb6-6038-4bbc-8aae-3713e26566d2","Type":"ContainerStarted","Data":"bb45610b05e23c34461cd6d02f4de0ebe48f60ac99be3fa49b55569ffe73cd73"} Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.507231 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-k7ct4"] Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.522992 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-kj4pg"] Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.524339 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.546881 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-kj4pg"] Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.710712 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-config\") pod \"dnsmasq-dns-5ccc8479f9-kj4pg\" (UID: \"1d6a066c-aa6f-4486-9d3e-298eb4b188ad\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.711102 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-kj4pg\" (UID: \"1d6a066c-aa6f-4486-9d3e-298eb4b188ad\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.711158 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czgxj\" (UniqueName: \"kubernetes.io/projected/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-kube-api-access-czgxj\") pod \"dnsmasq-dns-5ccc8479f9-kj4pg\" (UID: \"1d6a066c-aa6f-4486-9d3e-298eb4b188ad\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.811037 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-7c7nq"] Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.811928 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-kj4pg\" (UID: \"1d6a066c-aa6f-4486-9d3e-298eb4b188ad\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.812075 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czgxj\" (UniqueName: \"kubernetes.io/projected/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-kube-api-access-czgxj\") pod \"dnsmasq-dns-5ccc8479f9-kj4pg\" (UID: \"1d6a066c-aa6f-4486-9d3e-298eb4b188ad\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.812164 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-config\") pod \"dnsmasq-dns-5ccc8479f9-kj4pg\" (UID: \"1d6a066c-aa6f-4486-9d3e-298eb4b188ad\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.813172 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-config\") pod \"dnsmasq-dns-5ccc8479f9-kj4pg\" (UID: \"1d6a066c-aa6f-4486-9d3e-298eb4b188ad\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.815060 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-kj4pg\" (UID: \"1d6a066c-aa6f-4486-9d3e-298eb4b188ad\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.837587 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-kp4fh"] Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.838553 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czgxj\" (UniqueName: \"kubernetes.io/projected/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-kube-api-access-czgxj\") pod \"dnsmasq-dns-5ccc8479f9-kj4pg\" (UID: \"1d6a066c-aa6f-4486-9d3e-298eb4b188ad\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.840502 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.846291 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" Feb 27 16:41:44 crc kubenswrapper[4814]: I0227 16:41:44.856429 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-kp4fh"] Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.014788 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2368351c-7d12-4193-9d23-bac39e552df8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-kp4fh\" (UID: \"2368351c-7d12-4193-9d23-bac39e552df8\") " pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.015150 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2368351c-7d12-4193-9d23-bac39e552df8-config\") pod \"dnsmasq-dns-57d769cc4f-kp4fh\" (UID: \"2368351c-7d12-4193-9d23-bac39e552df8\") " pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.015168 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t8sc\" (UniqueName: \"kubernetes.io/projected/2368351c-7d12-4193-9d23-bac39e552df8-kube-api-access-4t8sc\") pod \"dnsmasq-dns-57d769cc4f-kp4fh\" (UID: \"2368351c-7d12-4193-9d23-bac39e552df8\") " pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.116694 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2368351c-7d12-4193-9d23-bac39e552df8-config\") pod \"dnsmasq-dns-57d769cc4f-kp4fh\" (UID: \"2368351c-7d12-4193-9d23-bac39e552df8\") " pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.116734 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t8sc\" (UniqueName: \"kubernetes.io/projected/2368351c-7d12-4193-9d23-bac39e552df8-kube-api-access-4t8sc\") pod \"dnsmasq-dns-57d769cc4f-kp4fh\" (UID: \"2368351c-7d12-4193-9d23-bac39e552df8\") " pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.116816 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2368351c-7d12-4193-9d23-bac39e552df8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-kp4fh\" (UID: \"2368351c-7d12-4193-9d23-bac39e552df8\") " pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.117607 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2368351c-7d12-4193-9d23-bac39e552df8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-kp4fh\" (UID: \"2368351c-7d12-4193-9d23-bac39e552df8\") " pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.117825 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2368351c-7d12-4193-9d23-bac39e552df8-config\") pod \"dnsmasq-dns-57d769cc4f-kp4fh\" (UID: \"2368351c-7d12-4193-9d23-bac39e552df8\") " pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.138304 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t8sc\" (UniqueName: \"kubernetes.io/projected/2368351c-7d12-4193-9d23-bac39e552df8-kube-api-access-4t8sc\") pod \"dnsmasq-dns-57d769cc4f-kp4fh\" (UID: \"2368351c-7d12-4193-9d23-bac39e552df8\") " pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.185762 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.347820 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-kj4pg"] Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.645574 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-kp4fh"] Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.674452 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.677234 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.678937 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.681557 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-8xdnh" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.682284 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.682603 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.682745 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.682913 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.686230 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.694139 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.759978 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" event={"ID":"1d6a066c-aa6f-4486-9d3e-298eb4b188ad","Type":"ContainerStarted","Data":"befa61bf7d512aa3537dba39092916b60904156b6160598970ca6383dc87fa0a"} Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.828318 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.828420 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.828469 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.828497 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.828532 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll9d5\" (UniqueName: \"kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-kube-api-access-ll9d5\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.828560 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.828583 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8bfc5fb5-c9ab-4108-915b-e6669b332e05-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.828600 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.828618 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8bfc5fb5-c9ab-4108-915b-e6669b332e05-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.828636 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.828655 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.929478 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.929550 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8bfc5fb5-c9ab-4108-915b-e6669b332e05-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.929572 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.929591 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8bfc5fb5-c9ab-4108-915b-e6669b332e05-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.929610 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.929627 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.929660 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.929691 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.929709 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.929732 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.929774 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll9d5\" (UniqueName: \"kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-kube-api-access-ll9d5\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.930735 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.930974 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.931013 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.931047 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.931129 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.934893 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.934926 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.938083 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.946006 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8bfc5fb5-c9ab-4108-915b-e6669b332e05-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.946146 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll9d5\" (UniqueName: \"kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-kube-api-access-ll9d5\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.947868 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8bfc5fb5-c9ab-4108-915b-e6669b332e05-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.957004 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:45 crc kubenswrapper[4814]: I0227 16:41:45.971031 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.002239 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.004111 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.008013 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.009561 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.009697 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.010059 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.010084 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.010345 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.010355 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.013212 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-dg2h7" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.134149 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.134215 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.134235 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.134268 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j42jm\" (UniqueName: \"kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-kube-api-access-j42jm\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.134306 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/58660097-6b23-4016-98b4-6a10978c0887-pod-info\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.134322 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-server-conf\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.134344 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.134364 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.134380 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/58660097-6b23-4016-98b4-6a10978c0887-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.134421 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-config-data\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.134439 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.237064 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-config-data\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.237121 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.237160 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.237198 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.237225 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.237272 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j42jm\" (UniqueName: \"kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-kube-api-access-j42jm\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.237312 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/58660097-6b23-4016-98b4-6a10978c0887-pod-info\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.237335 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-server-conf\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.237366 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.237399 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.237424 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/58660097-6b23-4016-98b4-6a10978c0887-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.237632 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.238053 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.238117 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.238781 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.239354 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-config-data\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.239526 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-server-conf\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.241328 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/58660097-6b23-4016-98b4-6a10978c0887-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.241631 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/58660097-6b23-4016-98b4-6a10978c0887-pod-info\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.241840 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.243542 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.255678 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j42jm\" (UniqueName: \"kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-kube-api-access-j42jm\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.264605 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " pod="openstack/rabbitmq-server-0" Feb 27 16:41:46 crc kubenswrapper[4814]: I0227 16:41:46.340435 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.033728 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.035569 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.044145 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-nsgdd" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.044768 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.045874 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.046334 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.049757 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.073931 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.152214 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/502232b7-9e95-4215-9dfb-08054b866199-config-data-generated\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.152346 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mjsk\" (UniqueName: \"kubernetes.io/projected/502232b7-9e95-4215-9dfb-08054b866199-kube-api-access-6mjsk\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.152389 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/502232b7-9e95-4215-9dfb-08054b866199-config-data-default\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.152416 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/502232b7-9e95-4215-9dfb-08054b866199-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.152457 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/502232b7-9e95-4215-9dfb-08054b866199-operator-scripts\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.152488 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/502232b7-9e95-4215-9dfb-08054b866199-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.152537 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.152553 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/502232b7-9e95-4215-9dfb-08054b866199-kolla-config\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.254622 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/502232b7-9e95-4215-9dfb-08054b866199-config-data-generated\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.254740 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mjsk\" (UniqueName: \"kubernetes.io/projected/502232b7-9e95-4215-9dfb-08054b866199-kube-api-access-6mjsk\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.254777 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/502232b7-9e95-4215-9dfb-08054b866199-config-data-default\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.254811 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/502232b7-9e95-4215-9dfb-08054b866199-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.254830 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/502232b7-9e95-4215-9dfb-08054b866199-operator-scripts\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.254849 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/502232b7-9e95-4215-9dfb-08054b866199-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.254927 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.254970 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/502232b7-9e95-4215-9dfb-08054b866199-kolla-config\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.255068 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/502232b7-9e95-4215-9dfb-08054b866199-config-data-generated\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.255450 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.256409 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/502232b7-9e95-4215-9dfb-08054b866199-kolla-config\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.260168 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/502232b7-9e95-4215-9dfb-08054b866199-config-data-default\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.260669 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/502232b7-9e95-4215-9dfb-08054b866199-operator-scripts\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.261982 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/502232b7-9e95-4215-9dfb-08054b866199-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.265022 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/502232b7-9e95-4215-9dfb-08054b866199-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.274410 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mjsk\" (UniqueName: \"kubernetes.io/projected/502232b7-9e95-4215-9dfb-08054b866199-kube-api-access-6mjsk\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.286056 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"502232b7-9e95-4215-9dfb-08054b866199\") " pod="openstack/openstack-galera-0" Feb 27 16:41:47 crc kubenswrapper[4814]: I0227 16:41:47.370213 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.360354 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.363582 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.367959 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-hfxb4" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.368018 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.368177 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.368228 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.374219 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.471761 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.472605 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czkv5\" (UniqueName: \"kubernetes.io/projected/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-kube-api-access-czkv5\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.472682 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.472721 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.472750 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.472785 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.472997 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.473028 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.576502 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.576562 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.576639 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.576675 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czkv5\" (UniqueName: \"kubernetes.io/projected/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-kube-api-access-czkv5\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.576703 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.576725 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.576743 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.576764 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.578696 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.581379 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.581980 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.582756 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.592427 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.593650 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.599239 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.606926 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.624139 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czkv5\" (UniqueName: \"kubernetes.io/projected/ea6bae63-9ffa-4ba7-a95f-e686e8d12e87-kube-api-access-czkv5\") pod \"openstack-cell1-galera-0\" (UID: \"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87\") " pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.680557 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.681650 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.684457 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.686323 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-9f6s6" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.686403 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.687733 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.695762 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.780284 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9glt\" (UniqueName: \"kubernetes.io/projected/414562e2-0c3b-4db9-bc71-78928f882398-kube-api-access-t9glt\") pod \"memcached-0\" (UID: \"414562e2-0c3b-4db9-bc71-78928f882398\") " pod="openstack/memcached-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.780885 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/414562e2-0c3b-4db9-bc71-78928f882398-kolla-config\") pod \"memcached-0\" (UID: \"414562e2-0c3b-4db9-bc71-78928f882398\") " pod="openstack/memcached-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.781091 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/414562e2-0c3b-4db9-bc71-78928f882398-config-data\") pod \"memcached-0\" (UID: \"414562e2-0c3b-4db9-bc71-78928f882398\") " pod="openstack/memcached-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.781180 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/414562e2-0c3b-4db9-bc71-78928f882398-memcached-tls-certs\") pod \"memcached-0\" (UID: \"414562e2-0c3b-4db9-bc71-78928f882398\") " pod="openstack/memcached-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.781205 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/414562e2-0c3b-4db9-bc71-78928f882398-combined-ca-bundle\") pod \"memcached-0\" (UID: \"414562e2-0c3b-4db9-bc71-78928f882398\") " pod="openstack/memcached-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.883204 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9glt\" (UniqueName: \"kubernetes.io/projected/414562e2-0c3b-4db9-bc71-78928f882398-kube-api-access-t9glt\") pod \"memcached-0\" (UID: \"414562e2-0c3b-4db9-bc71-78928f882398\") " pod="openstack/memcached-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.883308 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/414562e2-0c3b-4db9-bc71-78928f882398-kolla-config\") pod \"memcached-0\" (UID: \"414562e2-0c3b-4db9-bc71-78928f882398\") " pod="openstack/memcached-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.883382 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/414562e2-0c3b-4db9-bc71-78928f882398-config-data\") pod \"memcached-0\" (UID: \"414562e2-0c3b-4db9-bc71-78928f882398\") " pod="openstack/memcached-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.883410 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/414562e2-0c3b-4db9-bc71-78928f882398-memcached-tls-certs\") pod \"memcached-0\" (UID: \"414562e2-0c3b-4db9-bc71-78928f882398\") " pod="openstack/memcached-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.883429 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/414562e2-0c3b-4db9-bc71-78928f882398-combined-ca-bundle\") pod \"memcached-0\" (UID: \"414562e2-0c3b-4db9-bc71-78928f882398\") " pod="openstack/memcached-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.884896 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/414562e2-0c3b-4db9-bc71-78928f882398-config-data\") pod \"memcached-0\" (UID: \"414562e2-0c3b-4db9-bc71-78928f882398\") " pod="openstack/memcached-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.885231 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/414562e2-0c3b-4db9-bc71-78928f882398-kolla-config\") pod \"memcached-0\" (UID: \"414562e2-0c3b-4db9-bc71-78928f882398\") " pod="openstack/memcached-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.886817 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/414562e2-0c3b-4db9-bc71-78928f882398-combined-ca-bundle\") pod \"memcached-0\" (UID: \"414562e2-0c3b-4db9-bc71-78928f882398\") " pod="openstack/memcached-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.890052 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/414562e2-0c3b-4db9-bc71-78928f882398-memcached-tls-certs\") pod \"memcached-0\" (UID: \"414562e2-0c3b-4db9-bc71-78928f882398\") " pod="openstack/memcached-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.901952 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9glt\" (UniqueName: \"kubernetes.io/projected/414562e2-0c3b-4db9-bc71-78928f882398-kube-api-access-t9glt\") pod \"memcached-0\" (UID: \"414562e2-0c3b-4db9-bc71-78928f882398\") " pod="openstack/memcached-0" Feb 27 16:41:48 crc kubenswrapper[4814]: I0227 16:41:48.998176 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 27 16:41:49 crc kubenswrapper[4814]: W0227 16:41:49.438138 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2368351c_7d12_4193_9d23_bac39e552df8.slice/crio-7e2a1f93711ab34b5078550750fecb272933e68acd51289390a4d45e465b5826 WatchSource:0}: Error finding container 7e2a1f93711ab34b5078550750fecb272933e68acd51289390a4d45e465b5826: Status 404 returned error can't find the container with id 7e2a1f93711ab34b5078550750fecb272933e68acd51289390a4d45e465b5826 Feb 27 16:41:49 crc kubenswrapper[4814]: I0227 16:41:49.799219 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" event={"ID":"2368351c-7d12-4193-9d23-bac39e552df8","Type":"ContainerStarted","Data":"7e2a1f93711ab34b5078550750fecb272933e68acd51289390a4d45e465b5826"} Feb 27 16:41:51 crc kubenswrapper[4814]: I0227 16:41:51.077042 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 16:41:51 crc kubenswrapper[4814]: I0227 16:41:51.077970 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 16:41:51 crc kubenswrapper[4814]: I0227 16:41:51.080944 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-cpqd6" Feb 27 16:41:51 crc kubenswrapper[4814]: I0227 16:41:51.089397 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 16:41:51 crc kubenswrapper[4814]: I0227 16:41:51.225536 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnn94\" (UniqueName: \"kubernetes.io/projected/e0ea262c-91ce-44fd-b164-be82e6696e93-kube-api-access-pnn94\") pod \"kube-state-metrics-0\" (UID: \"e0ea262c-91ce-44fd-b164-be82e6696e93\") " pod="openstack/kube-state-metrics-0" Feb 27 16:41:51 crc kubenswrapper[4814]: I0227 16:41:51.327374 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnn94\" (UniqueName: \"kubernetes.io/projected/e0ea262c-91ce-44fd-b164-be82e6696e93-kube-api-access-pnn94\") pod \"kube-state-metrics-0\" (UID: \"e0ea262c-91ce-44fd-b164-be82e6696e93\") " pod="openstack/kube-state-metrics-0" Feb 27 16:41:51 crc kubenswrapper[4814]: I0227 16:41:51.367474 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnn94\" (UniqueName: \"kubernetes.io/projected/e0ea262c-91ce-44fd-b164-be82e6696e93-kube-api-access-pnn94\") pod \"kube-state-metrics-0\" (UID: \"e0ea262c-91ce-44fd-b164-be82e6696e93\") " pod="openstack/kube-state-metrics-0" Feb 27 16:41:51 crc kubenswrapper[4814]: I0227 16:41:51.396225 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 16:41:52 crc kubenswrapper[4814]: I0227 16:41:52.902851 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:41:52 crc kubenswrapper[4814]: I0227 16:41:52.903305 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.133628 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-9tkjl"] Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.134715 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.140052 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.140075 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.140227 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-tr8dl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.159000 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-kxqxw"] Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.160699 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9tkjl"] Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.160722 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-kxqxw"] Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.160812 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.277094 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62963d98-fb43-4fb1-939a-6f8e6b69cb16-scripts\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.277190 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/62963d98-fb43-4fb1-939a-6f8e6b69cb16-var-run\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.277415 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqcwn\" (UniqueName: \"kubernetes.io/projected/62963d98-fb43-4fb1-939a-6f8e6b69cb16-kube-api-access-pqcwn\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.277492 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c74238-2dde-4709-922a-9551da5fe8ae-combined-ca-bundle\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.277638 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/62963d98-fb43-4fb1-939a-6f8e6b69cb16-var-log\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.277688 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/56c74238-2dde-4709-922a-9551da5fe8ae-ovn-controller-tls-certs\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.277748 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/56c74238-2dde-4709-922a-9551da5fe8ae-var-run\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.277878 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/56c74238-2dde-4709-922a-9551da5fe8ae-scripts\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.277944 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/56c74238-2dde-4709-922a-9551da5fe8ae-var-run-ovn\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.277984 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l78s\" (UniqueName: \"kubernetes.io/projected/56c74238-2dde-4709-922a-9551da5fe8ae-kube-api-access-7l78s\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.278003 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/62963d98-fb43-4fb1-939a-6f8e6b69cb16-etc-ovs\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.278019 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/62963d98-fb43-4fb1-939a-6f8e6b69cb16-var-lib\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.278056 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/56c74238-2dde-4709-922a-9551da5fe8ae-var-log-ovn\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.378958 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/62963d98-fb43-4fb1-939a-6f8e6b69cb16-var-log\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379001 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/56c74238-2dde-4709-922a-9551da5fe8ae-ovn-controller-tls-certs\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379021 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/56c74238-2dde-4709-922a-9551da5fe8ae-var-run\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379044 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/56c74238-2dde-4709-922a-9551da5fe8ae-scripts\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379062 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/56c74238-2dde-4709-922a-9551da5fe8ae-var-run-ovn\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379081 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l78s\" (UniqueName: \"kubernetes.io/projected/56c74238-2dde-4709-922a-9551da5fe8ae-kube-api-access-7l78s\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379094 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/62963d98-fb43-4fb1-939a-6f8e6b69cb16-etc-ovs\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379111 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/62963d98-fb43-4fb1-939a-6f8e6b69cb16-var-lib\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379129 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/56c74238-2dde-4709-922a-9551da5fe8ae-var-log-ovn\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379167 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62963d98-fb43-4fb1-939a-6f8e6b69cb16-scripts\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379211 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/62963d98-fb43-4fb1-939a-6f8e6b69cb16-var-run\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379245 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqcwn\" (UniqueName: \"kubernetes.io/projected/62963d98-fb43-4fb1-939a-6f8e6b69cb16-kube-api-access-pqcwn\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379271 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c74238-2dde-4709-922a-9551da5fe8ae-combined-ca-bundle\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379482 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/62963d98-fb43-4fb1-939a-6f8e6b69cb16-var-log\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379547 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/56c74238-2dde-4709-922a-9551da5fe8ae-var-run\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379604 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/56c74238-2dde-4709-922a-9551da5fe8ae-var-run-ovn\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379653 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/62963d98-fb43-4fb1-939a-6f8e6b69cb16-etc-ovs\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379664 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/62963d98-fb43-4fb1-939a-6f8e6b69cb16-var-run\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379785 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/62963d98-fb43-4fb1-939a-6f8e6b69cb16-var-lib\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.379795 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/56c74238-2dde-4709-922a-9551da5fe8ae-var-log-ovn\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.381591 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62963d98-fb43-4fb1-939a-6f8e6b69cb16-scripts\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.381909 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/56c74238-2dde-4709-922a-9551da5fe8ae-scripts\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.386134 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c74238-2dde-4709-922a-9551da5fe8ae-combined-ca-bundle\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.389696 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/56c74238-2dde-4709-922a-9551da5fe8ae-ovn-controller-tls-certs\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.404129 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l78s\" (UniqueName: \"kubernetes.io/projected/56c74238-2dde-4709-922a-9551da5fe8ae-kube-api-access-7l78s\") pod \"ovn-controller-9tkjl\" (UID: \"56c74238-2dde-4709-922a-9551da5fe8ae\") " pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.404925 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqcwn\" (UniqueName: \"kubernetes.io/projected/62963d98-fb43-4fb1-939a-6f8e6b69cb16-kube-api-access-pqcwn\") pod \"ovn-controller-ovs-kxqxw\" (UID: \"62963d98-fb43-4fb1-939a-6f8e6b69cb16\") " pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.500743 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9tkjl" Feb 27 16:41:54 crc kubenswrapper[4814]: I0227 16:41:54.511097 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:41:55 crc kubenswrapper[4814]: I0227 16:41:55.984124 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 27 16:41:55 crc kubenswrapper[4814]: I0227 16:41:55.986414 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:55 crc kubenswrapper[4814]: I0227 16:41:55.989852 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 27 16:41:55 crc kubenswrapper[4814]: I0227 16:41:55.990315 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 27 16:41:55 crc kubenswrapper[4814]: I0227 16:41:55.990748 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 27 16:41:55 crc kubenswrapper[4814]: I0227 16:41:55.990747 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-qmzsn" Feb 27 16:41:55 crc kubenswrapper[4814]: I0227 16:41:55.991952 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 27 16:41:55 crc kubenswrapper[4814]: I0227 16:41:55.994629 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.114663 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bac59d25-01ed-44a7-9750-b654d1c7c631-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.114742 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8wcs\" (UniqueName: \"kubernetes.io/projected/bac59d25-01ed-44a7-9750-b654d1c7c631-kube-api-access-v8wcs\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.114792 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bac59d25-01ed-44a7-9750-b654d1c7c631-config\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.114827 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.114965 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bac59d25-01ed-44a7-9750-b654d1c7c631-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.115231 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac59d25-01ed-44a7-9750-b654d1c7c631-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.115347 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bac59d25-01ed-44a7-9750-b654d1c7c631-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.115593 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac59d25-01ed-44a7-9750-b654d1c7c631-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.217764 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bac59d25-01ed-44a7-9750-b654d1c7c631-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.217834 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac59d25-01ed-44a7-9750-b654d1c7c631-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.217868 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bac59d25-01ed-44a7-9750-b654d1c7c631-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.217938 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac59d25-01ed-44a7-9750-b654d1c7c631-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.217997 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bac59d25-01ed-44a7-9750-b654d1c7c631-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.218030 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8wcs\" (UniqueName: \"kubernetes.io/projected/bac59d25-01ed-44a7-9750-b654d1c7c631-kube-api-access-v8wcs\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.218056 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bac59d25-01ed-44a7-9750-b654d1c7c631-config\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.218073 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.218445 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.219394 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bac59d25-01ed-44a7-9750-b654d1c7c631-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.220120 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bac59d25-01ed-44a7-9750-b654d1c7c631-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.220275 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bac59d25-01ed-44a7-9750-b654d1c7c631-config\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.227185 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bac59d25-01ed-44a7-9750-b654d1c7c631-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.228564 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac59d25-01ed-44a7-9750-b654d1c7c631-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.241473 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac59d25-01ed-44a7-9750-b654d1c7c631-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.253006 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8wcs\" (UniqueName: \"kubernetes.io/projected/bac59d25-01ed-44a7-9750-b654d1c7c631-kube-api-access-v8wcs\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.258646 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"bac59d25-01ed-44a7-9750-b654d1c7c631\") " pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:56 crc kubenswrapper[4814]: I0227 16:41:56.314720 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 27 16:41:57 crc kubenswrapper[4814]: I0227 16:41:57.891789 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 27 16:41:57 crc kubenswrapper[4814]: I0227 16:41:57.893265 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:57 crc kubenswrapper[4814]: I0227 16:41:57.900721 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-wj4tx" Feb 27 16:41:57 crc kubenswrapper[4814]: I0227 16:41:57.900968 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 27 16:41:57 crc kubenswrapper[4814]: I0227 16:41:57.901111 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 27 16:41:57 crc kubenswrapper[4814]: I0227 16:41:57.901245 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 27 16:41:57 crc kubenswrapper[4814]: I0227 16:41:57.904596 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 27 16:41:57 crc kubenswrapper[4814]: I0227 16:41:57.952950 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/97ad2ba6-a67d-4926-93cd-1caf1c03e493-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:57 crc kubenswrapper[4814]: I0227 16:41:57.952999 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/97ad2ba6-a67d-4926-93cd-1caf1c03e493-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:57 crc kubenswrapper[4814]: I0227 16:41:57.953054 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:57 crc kubenswrapper[4814]: I0227 16:41:57.953315 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/97ad2ba6-a67d-4926-93cd-1caf1c03e493-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:57 crc kubenswrapper[4814]: I0227 16:41:57.953347 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97ad2ba6-a67d-4926-93cd-1caf1c03e493-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:57 crc kubenswrapper[4814]: I0227 16:41:57.953406 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97ad2ba6-a67d-4926-93cd-1caf1c03e493-config\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:57 crc kubenswrapper[4814]: I0227 16:41:57.953463 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97ad2ba6-a67d-4926-93cd-1caf1c03e493-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:57 crc kubenswrapper[4814]: I0227 16:41:57.953505 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pzxd\" (UniqueName: \"kubernetes.io/projected/97ad2ba6-a67d-4926-93cd-1caf1c03e493-kube-api-access-7pzxd\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.054686 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pzxd\" (UniqueName: \"kubernetes.io/projected/97ad2ba6-a67d-4926-93cd-1caf1c03e493-kube-api-access-7pzxd\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.054771 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/97ad2ba6-a67d-4926-93cd-1caf1c03e493-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.054806 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/97ad2ba6-a67d-4926-93cd-1caf1c03e493-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.054848 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.054883 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/97ad2ba6-a67d-4926-93cd-1caf1c03e493-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.054916 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97ad2ba6-a67d-4926-93cd-1caf1c03e493-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.054945 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97ad2ba6-a67d-4926-93cd-1caf1c03e493-config\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.054976 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97ad2ba6-a67d-4926-93cd-1caf1c03e493-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.056413 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.056700 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/97ad2ba6-a67d-4926-93cd-1caf1c03e493-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.057115 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97ad2ba6-a67d-4926-93cd-1caf1c03e493-config\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.058490 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97ad2ba6-a67d-4926-93cd-1caf1c03e493-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.060815 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/97ad2ba6-a67d-4926-93cd-1caf1c03e493-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.063618 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/97ad2ba6-a67d-4926-93cd-1caf1c03e493-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.064573 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97ad2ba6-a67d-4926-93cd-1caf1c03e493-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.074605 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pzxd\" (UniqueName: \"kubernetes.io/projected/97ad2ba6-a67d-4926-93cd-1caf1c03e493-kube-api-access-7pzxd\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.080547 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"97ad2ba6-a67d-4926-93cd-1caf1c03e493\") " pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: I0227 16:41:58.218363 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 27 16:41:58 crc kubenswrapper[4814]: E0227 16:41:58.494228 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 27 16:41:58 crc kubenswrapper[4814]: E0227 16:41:58.494398 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9h7zs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-k7ct4_openstack(6faabeb6-6038-4bbc-8aae-3713e26566d2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 16:41:58 crc kubenswrapper[4814]: E0227 16:41:58.498896 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-k7ct4" podUID="6faabeb6-6038-4bbc-8aae-3713e26566d2" Feb 27 16:41:58 crc kubenswrapper[4814]: E0227 16:41:58.540473 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 27 16:41:58 crc kubenswrapper[4814]: E0227 16:41:58.540653 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t84hk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-7c7nq_openstack(2183292c-2d97-4a2c-bc26-a46a0ffbf019): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 16:41:58 crc kubenswrapper[4814]: E0227 16:41:58.542034 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-7c7nq" podUID="2183292c-2d97-4a2c-bc26-a46a0ffbf019" Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.149236 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.176953 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.195328 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.519668 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.538922 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.568719 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.591426 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9tkjl"] Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.598771 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-7c7nq" Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.601283 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-k7ct4" Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.680420 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9h7zs\" (UniqueName: \"kubernetes.io/projected/6faabeb6-6038-4bbc-8aae-3713e26566d2-kube-api-access-9h7zs\") pod \"6faabeb6-6038-4bbc-8aae-3713e26566d2\" (UID: \"6faabeb6-6038-4bbc-8aae-3713e26566d2\") " Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.680532 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2183292c-2d97-4a2c-bc26-a46a0ffbf019-dns-svc\") pod \"2183292c-2d97-4a2c-bc26-a46a0ffbf019\" (UID: \"2183292c-2d97-4a2c-bc26-a46a0ffbf019\") " Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.680575 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t84hk\" (UniqueName: \"kubernetes.io/projected/2183292c-2d97-4a2c-bc26-a46a0ffbf019-kube-api-access-t84hk\") pod \"2183292c-2d97-4a2c-bc26-a46a0ffbf019\" (UID: \"2183292c-2d97-4a2c-bc26-a46a0ffbf019\") " Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.680614 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6faabeb6-6038-4bbc-8aae-3713e26566d2-config\") pod \"6faabeb6-6038-4bbc-8aae-3713e26566d2\" (UID: \"6faabeb6-6038-4bbc-8aae-3713e26566d2\") " Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.680653 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2183292c-2d97-4a2c-bc26-a46a0ffbf019-config\") pod \"2183292c-2d97-4a2c-bc26-a46a0ffbf019\" (UID: \"2183292c-2d97-4a2c-bc26-a46a0ffbf019\") " Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.681122 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2183292c-2d97-4a2c-bc26-a46a0ffbf019-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2183292c-2d97-4a2c-bc26-a46a0ffbf019" (UID: "2183292c-2d97-4a2c-bc26-a46a0ffbf019"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.681421 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2183292c-2d97-4a2c-bc26-a46a0ffbf019-config" (OuterVolumeSpecName: "config") pod "2183292c-2d97-4a2c-bc26-a46a0ffbf019" (UID: "2183292c-2d97-4a2c-bc26-a46a0ffbf019"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.681440 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6faabeb6-6038-4bbc-8aae-3713e26566d2-config" (OuterVolumeSpecName: "config") pod "6faabeb6-6038-4bbc-8aae-3713e26566d2" (UID: "6faabeb6-6038-4bbc-8aae-3713e26566d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.686227 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2183292c-2d97-4a2c-bc26-a46a0ffbf019-kube-api-access-t84hk" (OuterVolumeSpecName: "kube-api-access-t84hk") pod "2183292c-2d97-4a2c-bc26-a46a0ffbf019" (UID: "2183292c-2d97-4a2c-bc26-a46a0ffbf019"). InnerVolumeSpecName "kube-api-access-t84hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.686906 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6faabeb6-6038-4bbc-8aae-3713e26566d2-kube-api-access-9h7zs" (OuterVolumeSpecName: "kube-api-access-9h7zs") pod "6faabeb6-6038-4bbc-8aae-3713e26566d2" (UID: "6faabeb6-6038-4bbc-8aae-3713e26566d2"). InnerVolumeSpecName "kube-api-access-9h7zs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.708438 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.782150 4814 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2183292c-2d97-4a2c-bc26-a46a0ffbf019-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.782211 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t84hk\" (UniqueName: \"kubernetes.io/projected/2183292c-2d97-4a2c-bc26-a46a0ffbf019-kube-api-access-t84hk\") on node \"crc\" DevicePath \"\"" Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.782223 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6faabeb6-6038-4bbc-8aae-3713e26566d2-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.782232 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2183292c-2d97-4a2c-bc26-a46a0ffbf019-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.782241 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9h7zs\" (UniqueName: \"kubernetes.io/projected/6faabeb6-6038-4bbc-8aae-3713e26566d2-kube-api-access-9h7zs\") on node \"crc\" DevicePath \"\"" Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.885770 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9tkjl" event={"ID":"56c74238-2dde-4709-922a-9551da5fe8ae","Type":"ContainerStarted","Data":"20effa0c5fff40994017c50e00b056b6fb98a9c8663255eef167ba540022c070"} Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.887928 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-7c7nq" event={"ID":"2183292c-2d97-4a2c-bc26-a46a0ffbf019","Type":"ContainerDied","Data":"2a39b6fc3f3d8ff628b145c26562265d936ea6d5910b804d4c61514f85e99675"} Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.887958 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-7c7nq" Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.888979 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87","Type":"ContainerStarted","Data":"60c5fc7a20163e9aaa75d8d2e7569d2f10c4827802d6a49be694387d11983988"} Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.889739 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"414562e2-0c3b-4db9-bc71-78928f882398","Type":"ContainerStarted","Data":"28cd471ec6275691465cea8e7bb53464938230eb6b0f34fe7b564884b2de1fde"} Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.890817 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"97ad2ba6-a67d-4926-93cd-1caf1c03e493","Type":"ContainerStarted","Data":"637ed8db4e5c570dcd084e29c1d2bd25eea4937fa98d9556e7be3a4b5f9e0b77"} Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.892864 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8bfc5fb5-c9ab-4108-915b-e6669b332e05","Type":"ContainerStarted","Data":"fd5105f23466970276cfc746c6b0fd5d2ec801f03303da64073e4d930913f7eb"} Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.894531 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"502232b7-9e95-4215-9dfb-08054b866199","Type":"ContainerStarted","Data":"1f12e7e6f7ca4304f2535594ddd692fc634c773d3f30e730f5bd7b0feb738e64"} Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.895463 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e0ea262c-91ce-44fd-b164-be82e6696e93","Type":"ContainerStarted","Data":"3276ce6178b3b4ca62756118b2280a92a9d07654f250dbe152af22e69dca4af6"} Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.896659 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-k7ct4" event={"ID":"6faabeb6-6038-4bbc-8aae-3713e26566d2","Type":"ContainerDied","Data":"bb45610b05e23c34461cd6d02f4de0ebe48f60ac99be3fa49b55569ffe73cd73"} Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.896678 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-k7ct4" Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.897819 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"58660097-6b23-4016-98b4-6a10978c0887","Type":"ContainerStarted","Data":"d12bf0f65745d1088e3f5b639958b7edfd15f8e0f3f1a158a85811b8e0a9c149"} Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.900515 4814 generic.go:334] "Generic (PLEG): container finished" podID="1d6a066c-aa6f-4486-9d3e-298eb4b188ad" containerID="e060285c5d70a9ab6c0995b4e43f81ee6f474a18192f9a691b33d69c2d0bb995" exitCode=0 Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.900583 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" event={"ID":"1d6a066c-aa6f-4486-9d3e-298eb4b188ad","Type":"ContainerDied","Data":"e060285c5d70a9ab6c0995b4e43f81ee6f474a18192f9a691b33d69c2d0bb995"} Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.902423 4814 generic.go:334] "Generic (PLEG): container finished" podID="2368351c-7d12-4193-9d23-bac39e552df8" containerID="d27f66b53b70af6a019c1e53651b3c4e5acfe515cb3292df205015798e737991" exitCode=0 Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.902456 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" event={"ID":"2368351c-7d12-4193-9d23-bac39e552df8","Type":"ContainerDied","Data":"d27f66b53b70af6a019c1e53651b3c4e5acfe515cb3292df205015798e737991"} Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.972077 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-7c7nq"] Feb 27 16:41:59 crc kubenswrapper[4814]: I0227 16:41:59.978920 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-7c7nq"] Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.035357 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-k7ct4"] Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.064779 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-k7ct4"] Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.128831 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536842-hvtr9"] Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.134623 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536842-hvtr9" Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.136132 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.136213 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.138585 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.138613 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536842-hvtr9"] Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.188470 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwfdg\" (UniqueName: \"kubernetes.io/projected/c2edd190-afdf-4d53-a42e-c8a41e4a8853-kube-api-access-zwfdg\") pod \"auto-csr-approver-29536842-hvtr9\" (UID: \"c2edd190-afdf-4d53-a42e-c8a41e4a8853\") " pod="openshift-infra/auto-csr-approver-29536842-hvtr9" Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.289589 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwfdg\" (UniqueName: \"kubernetes.io/projected/c2edd190-afdf-4d53-a42e-c8a41e4a8853-kube-api-access-zwfdg\") pod \"auto-csr-approver-29536842-hvtr9\" (UID: \"c2edd190-afdf-4d53-a42e-c8a41e4a8853\") " pod="openshift-infra/auto-csr-approver-29536842-hvtr9" Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.308628 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwfdg\" (UniqueName: \"kubernetes.io/projected/c2edd190-afdf-4d53-a42e-c8a41e4a8853-kube-api-access-zwfdg\") pod \"auto-csr-approver-29536842-hvtr9\" (UID: \"c2edd190-afdf-4d53-a42e-c8a41e4a8853\") " pod="openshift-infra/auto-csr-approver-29536842-hvtr9" Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.458877 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.468980 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536842-hvtr9" Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.500601 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2183292c-2d97-4a2c-bc26-a46a0ffbf019" path="/var/lib/kubelet/pods/2183292c-2d97-4a2c-bc26-a46a0ffbf019/volumes" Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.500992 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6faabeb6-6038-4bbc-8aae-3713e26566d2" path="/var/lib/kubelet/pods/6faabeb6-6038-4bbc-8aae-3713e26566d2/volumes" Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.559553 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-kxqxw"] Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.912692 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kxqxw" event={"ID":"62963d98-fb43-4fb1-939a-6f8e6b69cb16","Type":"ContainerStarted","Data":"1e24ae9397c0208c96ea509c90408f295874a8d224161fee2140dc2b95c5e0f3"} Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.914899 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bac59d25-01ed-44a7-9750-b654d1c7c631","Type":"ContainerStarted","Data":"a146ebe76d2032af5742573152a9955b34b6523bc5ebc86157a97ae16ae2cf7c"} Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.917062 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" event={"ID":"1d6a066c-aa6f-4486-9d3e-298eb4b188ad","Type":"ContainerStarted","Data":"1527eade6f48d16b663c5911afd43b1b791e7582dd57d1a7671a03b14b7a2937"} Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.917200 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.920403 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" event={"ID":"2368351c-7d12-4193-9d23-bac39e552df8","Type":"ContainerStarted","Data":"c74323d36faf3ba14bc0582dd74bf9c49679b5c81d83a9dac533100e03ed07f4"} Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.921636 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.940943 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" podStartSLOduration=3.636152497 podStartE2EDuration="16.940920027s" podCreationTimestamp="2026-02-27 16:41:44 +0000 UTC" firstStartedPulling="2026-02-27 16:41:45.35296512 +0000 UTC m=+1117.805589950" lastFinishedPulling="2026-02-27 16:41:58.65773265 +0000 UTC m=+1131.110357480" observedRunningTime="2026-02-27 16:42:00.935122971 +0000 UTC m=+1133.387747801" watchObservedRunningTime="2026-02-27 16:42:00.940920027 +0000 UTC m=+1133.393544857" Feb 27 16:42:00 crc kubenswrapper[4814]: I0227 16:42:00.964752 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" podStartSLOduration=7.760675745 podStartE2EDuration="16.96473054s" podCreationTimestamp="2026-02-27 16:41:44 +0000 UTC" firstStartedPulling="2026-02-27 16:41:49.440541696 +0000 UTC m=+1121.893166566" lastFinishedPulling="2026-02-27 16:41:58.644596541 +0000 UTC m=+1131.097221361" observedRunningTime="2026-02-27 16:42:00.956187341 +0000 UTC m=+1133.408812171" watchObservedRunningTime="2026-02-27 16:42:00.96473054 +0000 UTC m=+1133.417355380" Feb 27 16:42:05 crc kubenswrapper[4814]: I0227 16:42:05.187442 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" Feb 27 16:42:05 crc kubenswrapper[4814]: I0227 16:42:05.236164 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-kj4pg"] Feb 27 16:42:05 crc kubenswrapper[4814]: I0227 16:42:05.236384 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" podUID="1d6a066c-aa6f-4486-9d3e-298eb4b188ad" containerName="dnsmasq-dns" containerID="cri-o://1527eade6f48d16b663c5911afd43b1b791e7582dd57d1a7671a03b14b7a2937" gracePeriod=10 Feb 27 16:42:05 crc kubenswrapper[4814]: I0227 16:42:05.241468 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" Feb 27 16:42:05 crc kubenswrapper[4814]: I0227 16:42:05.963366 4814 generic.go:334] "Generic (PLEG): container finished" podID="1d6a066c-aa6f-4486-9d3e-298eb4b188ad" containerID="1527eade6f48d16b663c5911afd43b1b791e7582dd57d1a7671a03b14b7a2937" exitCode=0 Feb 27 16:42:05 crc kubenswrapper[4814]: I0227 16:42:05.963399 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" event={"ID":"1d6a066c-aa6f-4486-9d3e-298eb4b188ad","Type":"ContainerDied","Data":"1527eade6f48d16b663c5911afd43b1b791e7582dd57d1a7671a03b14b7a2937"} Feb 27 16:42:07 crc kubenswrapper[4814]: I0227 16:42:07.941336 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" Feb 27 16:42:07 crc kubenswrapper[4814]: I0227 16:42:07.995420 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" event={"ID":"1d6a066c-aa6f-4486-9d3e-298eb4b188ad","Type":"ContainerDied","Data":"befa61bf7d512aa3537dba39092916b60904156b6160598970ca6383dc87fa0a"} Feb 27 16:42:07 crc kubenswrapper[4814]: I0227 16:42:07.995481 4814 scope.go:117] "RemoveContainer" containerID="1527eade6f48d16b663c5911afd43b1b791e7582dd57d1a7671a03b14b7a2937" Feb 27 16:42:07 crc kubenswrapper[4814]: I0227 16:42:07.995835 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-kj4pg" Feb 27 16:42:08 crc kubenswrapper[4814]: I0227 16:42:08.048888 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czgxj\" (UniqueName: \"kubernetes.io/projected/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-kube-api-access-czgxj\") pod \"1d6a066c-aa6f-4486-9d3e-298eb4b188ad\" (UID: \"1d6a066c-aa6f-4486-9d3e-298eb4b188ad\") " Feb 27 16:42:08 crc kubenswrapper[4814]: I0227 16:42:08.048943 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-dns-svc\") pod \"1d6a066c-aa6f-4486-9d3e-298eb4b188ad\" (UID: \"1d6a066c-aa6f-4486-9d3e-298eb4b188ad\") " Feb 27 16:42:08 crc kubenswrapper[4814]: I0227 16:42:08.048999 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-config\") pod \"1d6a066c-aa6f-4486-9d3e-298eb4b188ad\" (UID: \"1d6a066c-aa6f-4486-9d3e-298eb4b188ad\") " Feb 27 16:42:08 crc kubenswrapper[4814]: I0227 16:42:08.056691 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-kube-api-access-czgxj" (OuterVolumeSpecName: "kube-api-access-czgxj") pod "1d6a066c-aa6f-4486-9d3e-298eb4b188ad" (UID: "1d6a066c-aa6f-4486-9d3e-298eb4b188ad"). InnerVolumeSpecName "kube-api-access-czgxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:42:08 crc kubenswrapper[4814]: I0227 16:42:08.095851 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-config" (OuterVolumeSpecName: "config") pod "1d6a066c-aa6f-4486-9d3e-298eb4b188ad" (UID: "1d6a066c-aa6f-4486-9d3e-298eb4b188ad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:08 crc kubenswrapper[4814]: I0227 16:42:08.114035 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1d6a066c-aa6f-4486-9d3e-298eb4b188ad" (UID: "1d6a066c-aa6f-4486-9d3e-298eb4b188ad"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:08 crc kubenswrapper[4814]: I0227 16:42:08.150860 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czgxj\" (UniqueName: \"kubernetes.io/projected/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-kube-api-access-czgxj\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:08 crc kubenswrapper[4814]: I0227 16:42:08.150894 4814 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:08 crc kubenswrapper[4814]: I0227 16:42:08.150905 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d6a066c-aa6f-4486-9d3e-298eb4b188ad-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:08 crc kubenswrapper[4814]: I0227 16:42:08.346419 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-kj4pg"] Feb 27 16:42:08 crc kubenswrapper[4814]: I0227 16:42:08.351183 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-kj4pg"] Feb 27 16:42:08 crc kubenswrapper[4814]: I0227 16:42:08.364952 4814 scope.go:117] "RemoveContainer" containerID="e060285c5d70a9ab6c0995b4e43f81ee6f474a18192f9a691b33d69c2d0bb995" Feb 27 16:42:08 crc kubenswrapper[4814]: I0227 16:42:08.395072 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536842-hvtr9"] Feb 27 16:42:08 crc kubenswrapper[4814]: I0227 16:42:08.504169 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d6a066c-aa6f-4486-9d3e-298eb4b188ad" path="/var/lib/kubelet/pods/1d6a066c-aa6f-4486-9d3e-298eb4b188ad/volumes" Feb 27 16:42:08 crc kubenswrapper[4814]: W0227 16:42:08.924411 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2edd190_afdf_4d53_a42e_c8a41e4a8853.slice/crio-538191741d66e6cee2c8276688045b4cf7e92b5347b0f0d6e7493387048b1e5d WatchSource:0}: Error finding container 538191741d66e6cee2c8276688045b4cf7e92b5347b0f0d6e7493387048b1e5d: Status 404 returned error can't find the container with id 538191741d66e6cee2c8276688045b4cf7e92b5347b0f0d6e7493387048b1e5d Feb 27 16:42:09 crc kubenswrapper[4814]: I0227 16:42:09.007470 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536842-hvtr9" event={"ID":"c2edd190-afdf-4d53-a42e-c8a41e4a8853","Type":"ContainerStarted","Data":"538191741d66e6cee2c8276688045b4cf7e92b5347b0f0d6e7493387048b1e5d"} Feb 27 16:42:10 crc kubenswrapper[4814]: I0227 16:42:10.017875 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87","Type":"ContainerStarted","Data":"f284ada36670bb058291efd3d04fc76d860bf0310a629674309a6b78f9b24fb2"} Feb 27 16:42:10 crc kubenswrapper[4814]: I0227 16:42:10.021690 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e0ea262c-91ce-44fd-b164-be82e6696e93","Type":"ContainerStarted","Data":"91632a104e79e38caed145351cf91b5ac043517bc37f9849b25e7a6cbb580785"} Feb 27 16:42:10 crc kubenswrapper[4814]: I0227 16:42:10.022092 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 27 16:42:10 crc kubenswrapper[4814]: I0227 16:42:10.024415 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"414562e2-0c3b-4db9-bc71-78928f882398","Type":"ContainerStarted","Data":"9cdc686f09c037a9f0720a1b78a51c7c93990445929706a90c051482f4b91dac"} Feb 27 16:42:10 crc kubenswrapper[4814]: I0227 16:42:10.024883 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 27 16:42:10 crc kubenswrapper[4814]: I0227 16:42:10.026842 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"97ad2ba6-a67d-4926-93cd-1caf1c03e493","Type":"ContainerStarted","Data":"b8393e3bea46a5b3315ec8defe8201929b46f840e74fec7a799c0e0416147916"} Feb 27 16:42:10 crc kubenswrapper[4814]: I0227 16:42:10.030841 4814 generic.go:334] "Generic (PLEG): container finished" podID="62963d98-fb43-4fb1-939a-6f8e6b69cb16" containerID="38d3507121f2d2de45f138c41f22fd1848ba2d4fb00cbd7521085a6e9b028448" exitCode=0 Feb 27 16:42:10 crc kubenswrapper[4814]: I0227 16:42:10.030976 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kxqxw" event={"ID":"62963d98-fb43-4fb1-939a-6f8e6b69cb16","Type":"ContainerDied","Data":"38d3507121f2d2de45f138c41f22fd1848ba2d4fb00cbd7521085a6e9b028448"} Feb 27 16:42:10 crc kubenswrapper[4814]: I0227 16:42:10.034902 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bac59d25-01ed-44a7-9750-b654d1c7c631","Type":"ContainerStarted","Data":"9e1da71009df7eb24fd1ff86880d8b41b0a44ab8222119254fad0880b8a91e9b"} Feb 27 16:42:10 crc kubenswrapper[4814]: I0227 16:42:10.037486 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"502232b7-9e95-4215-9dfb-08054b866199","Type":"ContainerStarted","Data":"c8c483c55dbaf45a63baba8349da0ae5e012dbd3c584de954b5b6c10d35dd3a0"} Feb 27 16:42:10 crc kubenswrapper[4814]: I0227 16:42:10.038693 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9tkjl" event={"ID":"56c74238-2dde-4709-922a-9551da5fe8ae","Type":"ContainerStarted","Data":"4fbcee0ce2492a3194b73c92aafdb5b60a61e3347c1ab4a8b88dfd0c62233b57"} Feb 27 16:42:10 crc kubenswrapper[4814]: I0227 16:42:10.039100 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-9tkjl" Feb 27 16:42:10 crc kubenswrapper[4814]: I0227 16:42:10.117628 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=13.965038286 podStartE2EDuration="22.117611061s" podCreationTimestamp="2026-02-27 16:41:48 +0000 UTC" firstStartedPulling="2026-02-27 16:41:59.205377311 +0000 UTC m=+1131.658002141" lastFinishedPulling="2026-02-27 16:42:07.357950086 +0000 UTC m=+1139.810574916" observedRunningTime="2026-02-27 16:42:10.089064053 +0000 UTC m=+1142.541688883" watchObservedRunningTime="2026-02-27 16:42:10.117611061 +0000 UTC m=+1142.570235881" Feb 27 16:42:10 crc kubenswrapper[4814]: I0227 16:42:10.120361 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-9tkjl" podStartSLOduration=7.521761386 podStartE2EDuration="16.120351164s" podCreationTimestamp="2026-02-27 16:41:54 +0000 UTC" firstStartedPulling="2026-02-27 16:41:59.624012691 +0000 UTC m=+1132.076637531" lastFinishedPulling="2026-02-27 16:42:08.222602479 +0000 UTC m=+1140.675227309" observedRunningTime="2026-02-27 16:42:10.109794733 +0000 UTC m=+1142.562419563" watchObservedRunningTime="2026-02-27 16:42:10.120351164 +0000 UTC m=+1142.572975994" Feb 27 16:42:10 crc kubenswrapper[4814]: I0227 16:42:10.146455 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=9.66917697 podStartE2EDuration="19.146435287s" podCreationTimestamp="2026-02-27 16:41:51 +0000 UTC" firstStartedPulling="2026-02-27 16:41:59.590837783 +0000 UTC m=+1132.043462613" lastFinishedPulling="2026-02-27 16:42:09.0680961 +0000 UTC m=+1141.520720930" observedRunningTime="2026-02-27 16:42:10.138120734 +0000 UTC m=+1142.590745614" watchObservedRunningTime="2026-02-27 16:42:10.146435287 +0000 UTC m=+1142.599060107" Feb 27 16:42:11 crc kubenswrapper[4814]: I0227 16:42:11.047943 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kxqxw" event={"ID":"62963d98-fb43-4fb1-939a-6f8e6b69cb16","Type":"ContainerStarted","Data":"bbc0ffe782a198b520172bf4f4216420ff5f3849b4e55e4a709661878a7d2344"} Feb 27 16:42:11 crc kubenswrapper[4814]: I0227 16:42:11.048185 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kxqxw" event={"ID":"62963d98-fb43-4fb1-939a-6f8e6b69cb16","Type":"ContainerStarted","Data":"0aa8953aabe1a0df842cfec0316139981acdef778c4dc129214d44bac2ecfcba"} Feb 27 16:42:11 crc kubenswrapper[4814]: I0227 16:42:11.048507 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:42:11 crc kubenswrapper[4814]: I0227 16:42:11.048545 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:42:11 crc kubenswrapper[4814]: I0227 16:42:11.051325 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8bfc5fb5-c9ab-4108-915b-e6669b332e05","Type":"ContainerStarted","Data":"2c129c7f4218bd10e27a0ca4ecd6c565a206b1a457efa3f4a5f5c065833f539d"} Feb 27 16:42:11 crc kubenswrapper[4814]: I0227 16:42:11.055891 4814 generic.go:334] "Generic (PLEG): container finished" podID="c2edd190-afdf-4d53-a42e-c8a41e4a8853" containerID="e02cda0f2dd0eec50057588dc8a38645b05d52a2e967dd1110002b8ccbab6273" exitCode=0 Feb 27 16:42:11 crc kubenswrapper[4814]: I0227 16:42:11.056036 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536842-hvtr9" event={"ID":"c2edd190-afdf-4d53-a42e-c8a41e4a8853","Type":"ContainerDied","Data":"e02cda0f2dd0eec50057588dc8a38645b05d52a2e967dd1110002b8ccbab6273"} Feb 27 16:42:11 crc kubenswrapper[4814]: I0227 16:42:11.059362 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"58660097-6b23-4016-98b4-6a10978c0887","Type":"ContainerStarted","Data":"3901d5b87aa0457909d2bb6c3ba5880bef5442d1430009b012abf06b49f82fba"} Feb 27 16:42:11 crc kubenswrapper[4814]: I0227 16:42:11.072235 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-kxqxw" podStartSLOduration=9.738820684 podStartE2EDuration="17.072217907s" podCreationTimestamp="2026-02-27 16:41:54 +0000 UTC" firstStartedPulling="2026-02-27 16:42:00.794802177 +0000 UTC m=+1133.247427007" lastFinishedPulling="2026-02-27 16:42:08.1281994 +0000 UTC m=+1140.580824230" observedRunningTime="2026-02-27 16:42:11.066034189 +0000 UTC m=+1143.518659019" watchObservedRunningTime="2026-02-27 16:42:11.072217907 +0000 UTC m=+1143.524842737" Feb 27 16:42:13 crc kubenswrapper[4814]: I0227 16:42:13.311748 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536842-hvtr9" Feb 27 16:42:13 crc kubenswrapper[4814]: I0227 16:42:13.456860 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwfdg\" (UniqueName: \"kubernetes.io/projected/c2edd190-afdf-4d53-a42e-c8a41e4a8853-kube-api-access-zwfdg\") pod \"c2edd190-afdf-4d53-a42e-c8a41e4a8853\" (UID: \"c2edd190-afdf-4d53-a42e-c8a41e4a8853\") " Feb 27 16:42:13 crc kubenswrapper[4814]: I0227 16:42:13.463083 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2edd190-afdf-4d53-a42e-c8a41e4a8853-kube-api-access-zwfdg" (OuterVolumeSpecName: "kube-api-access-zwfdg") pod "c2edd190-afdf-4d53-a42e-c8a41e4a8853" (UID: "c2edd190-afdf-4d53-a42e-c8a41e4a8853"). InnerVolumeSpecName "kube-api-access-zwfdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:42:13 crc kubenswrapper[4814]: I0227 16:42:13.558590 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwfdg\" (UniqueName: \"kubernetes.io/projected/c2edd190-afdf-4d53-a42e-c8a41e4a8853-kube-api-access-zwfdg\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:14 crc kubenswrapper[4814]: I0227 16:42:14.088737 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"97ad2ba6-a67d-4926-93cd-1caf1c03e493","Type":"ContainerStarted","Data":"788fb0e94298566b7ce8d5a398d64e3ba9f6906a53200fd9c258361c47a42f6b"} Feb 27 16:42:14 crc kubenswrapper[4814]: I0227 16:42:14.091531 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bac59d25-01ed-44a7-9750-b654d1c7c631","Type":"ContainerStarted","Data":"7dc585e22767421c70ebe5f99567fc03e9b0b7bdd8dbb3c38877b0c19674dbdb"} Feb 27 16:42:14 crc kubenswrapper[4814]: I0227 16:42:14.093865 4814 generic.go:334] "Generic (PLEG): container finished" podID="502232b7-9e95-4215-9dfb-08054b866199" containerID="c8c483c55dbaf45a63baba8349da0ae5e012dbd3c584de954b5b6c10d35dd3a0" exitCode=0 Feb 27 16:42:14 crc kubenswrapper[4814]: I0227 16:42:14.093934 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"502232b7-9e95-4215-9dfb-08054b866199","Type":"ContainerDied","Data":"c8c483c55dbaf45a63baba8349da0ae5e012dbd3c584de954b5b6c10d35dd3a0"} Feb 27 16:42:14 crc kubenswrapper[4814]: I0227 16:42:14.096177 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536842-hvtr9" event={"ID":"c2edd190-afdf-4d53-a42e-c8a41e4a8853","Type":"ContainerDied","Data":"538191741d66e6cee2c8276688045b4cf7e92b5347b0f0d6e7493387048b1e5d"} Feb 27 16:42:14 crc kubenswrapper[4814]: I0227 16:42:14.096202 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="538191741d66e6cee2c8276688045b4cf7e92b5347b0f0d6e7493387048b1e5d" Feb 27 16:42:14 crc kubenswrapper[4814]: I0227 16:42:14.096241 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536842-hvtr9" Feb 27 16:42:14 crc kubenswrapper[4814]: I0227 16:42:14.120108 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=3.915706714 podStartE2EDuration="18.12008847s" podCreationTimestamp="2026-02-27 16:41:56 +0000 UTC" firstStartedPulling="2026-02-27 16:41:59.715422629 +0000 UTC m=+1132.168047459" lastFinishedPulling="2026-02-27 16:42:13.919804385 +0000 UTC m=+1146.372429215" observedRunningTime="2026-02-27 16:42:14.11612474 +0000 UTC m=+1146.568749610" watchObservedRunningTime="2026-02-27 16:42:14.12008847 +0000 UTC m=+1146.572713300" Feb 27 16:42:14 crc kubenswrapper[4814]: I0227 16:42:14.149242 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=7.060651094 podStartE2EDuration="20.149218265s" podCreationTimestamp="2026-02-27 16:41:54 +0000 UTC" firstStartedPulling="2026-02-27 16:42:00.802615204 +0000 UTC m=+1133.255240034" lastFinishedPulling="2026-02-27 16:42:13.891182375 +0000 UTC m=+1146.343807205" observedRunningTime="2026-02-27 16:42:14.144133501 +0000 UTC m=+1146.596758371" watchObservedRunningTime="2026-02-27 16:42:14.149218265 +0000 UTC m=+1146.601843135" Feb 27 16:42:14 crc kubenswrapper[4814]: I0227 16:42:14.315185 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 27 16:42:14 crc kubenswrapper[4814]: I0227 16:42:14.377095 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 27 16:42:14 crc kubenswrapper[4814]: I0227 16:42:14.397824 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536836-7pdzj"] Feb 27 16:42:14 crc kubenswrapper[4814]: I0227 16:42:14.408711 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536836-7pdzj"] Feb 27 16:42:14 crc kubenswrapper[4814]: I0227 16:42:14.494772 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91" path="/var/lib/kubelet/pods/83a6bfb1-aeed-4ad2-a22c-ae30c36fbb91/volumes" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.107513 4814 generic.go:334] "Generic (PLEG): container finished" podID="ea6bae63-9ffa-4ba7-a95f-e686e8d12e87" containerID="f284ada36670bb058291efd3d04fc76d860bf0310a629674309a6b78f9b24fb2" exitCode=0 Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.107592 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87","Type":"ContainerDied","Data":"f284ada36670bb058291efd3d04fc76d860bf0310a629674309a6b78f9b24fb2"} Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.110277 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"502232b7-9e95-4215-9dfb-08054b866199","Type":"ContainerStarted","Data":"718a1397ecf3e8f1910c42052d992aaff697790de82c8513aaab229b2de2beb9"} Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.111075 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.177128 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=21.361038563 podStartE2EDuration="30.177101869s" podCreationTimestamp="2026-02-27 16:41:45 +0000 UTC" firstStartedPulling="2026-02-27 16:41:59.538206924 +0000 UTC m=+1131.990831754" lastFinishedPulling="2026-02-27 16:42:08.35427021 +0000 UTC m=+1140.806895060" observedRunningTime="2026-02-27 16:42:15.164190176 +0000 UTC m=+1147.616815046" watchObservedRunningTime="2026-02-27 16:42:15.177101869 +0000 UTC m=+1147.629726739" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.181520 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.491782 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-h9lfc"] Feb 27 16:42:15 crc kubenswrapper[4814]: E0227 16:42:15.492194 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d6a066c-aa6f-4486-9d3e-298eb4b188ad" containerName="init" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.492214 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d6a066c-aa6f-4486-9d3e-298eb4b188ad" containerName="init" Feb 27 16:42:15 crc kubenswrapper[4814]: E0227 16:42:15.492267 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d6a066c-aa6f-4486-9d3e-298eb4b188ad" containerName="dnsmasq-dns" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.492294 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d6a066c-aa6f-4486-9d3e-298eb4b188ad" containerName="dnsmasq-dns" Feb 27 16:42:15 crc kubenswrapper[4814]: E0227 16:42:15.492310 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2edd190-afdf-4d53-a42e-c8a41e4a8853" containerName="oc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.492320 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2edd190-afdf-4d53-a42e-c8a41e4a8853" containerName="oc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.492526 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d6a066c-aa6f-4486-9d3e-298eb4b188ad" containerName="dnsmasq-dns" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.492544 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2edd190-afdf-4d53-a42e-c8a41e4a8853" containerName="oc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.493560 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.497817 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.509142 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-h9lfc"] Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.552502 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-hwlpr"] Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.553544 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.556383 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.569808 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-hwlpr"] Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.595572 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-h9lfc\" (UID: \"a975048b-7b28-4784-8ae2-cadacba59fdc\") " pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.595669 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv94f\" (UniqueName: \"kubernetes.io/projected/a975048b-7b28-4784-8ae2-cadacba59fdc-kube-api-access-lv94f\") pod \"dnsmasq-dns-7f896c8c65-h9lfc\" (UID: \"a975048b-7b28-4784-8ae2-cadacba59fdc\") " pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.595697 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-config\") pod \"dnsmasq-dns-7f896c8c65-h9lfc\" (UID: \"a975048b-7b28-4784-8ae2-cadacba59fdc\") " pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.595765 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-h9lfc\" (UID: \"a975048b-7b28-4784-8ae2-cadacba59fdc\") " pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.697469 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3a71748-678d-498b-8a4b-6bf26679c1cc-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.697539 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv94f\" (UniqueName: \"kubernetes.io/projected/a975048b-7b28-4784-8ae2-cadacba59fdc-kube-api-access-lv94f\") pod \"dnsmasq-dns-7f896c8c65-h9lfc\" (UID: \"a975048b-7b28-4784-8ae2-cadacba59fdc\") " pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.697571 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b3a71748-678d-498b-8a4b-6bf26679c1cc-ovs-rundir\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.697600 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-config\") pod \"dnsmasq-dns-7f896c8c65-h9lfc\" (UID: \"a975048b-7b28-4784-8ae2-cadacba59fdc\") " pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.697648 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-h9lfc\" (UID: \"a975048b-7b28-4784-8ae2-cadacba59fdc\") " pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.697686 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdvfq\" (UniqueName: \"kubernetes.io/projected/b3a71748-678d-498b-8a4b-6bf26679c1cc-kube-api-access-zdvfq\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.697718 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3a71748-678d-498b-8a4b-6bf26679c1cc-combined-ca-bundle\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.697779 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3a71748-678d-498b-8a4b-6bf26679c1cc-config\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.697805 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-h9lfc\" (UID: \"a975048b-7b28-4784-8ae2-cadacba59fdc\") " pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.697827 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b3a71748-678d-498b-8a4b-6bf26679c1cc-ovn-rundir\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.698416 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-config\") pod \"dnsmasq-dns-7f896c8c65-h9lfc\" (UID: \"a975048b-7b28-4784-8ae2-cadacba59fdc\") " pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.698482 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-h9lfc\" (UID: \"a975048b-7b28-4784-8ae2-cadacba59fdc\") " pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.698804 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-h9lfc\" (UID: \"a975048b-7b28-4784-8ae2-cadacba59fdc\") " pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.713945 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv94f\" (UniqueName: \"kubernetes.io/projected/a975048b-7b28-4784-8ae2-cadacba59fdc-kube-api-access-lv94f\") pod \"dnsmasq-dns-7f896c8c65-h9lfc\" (UID: \"a975048b-7b28-4784-8ae2-cadacba59fdc\") " pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.799218 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b3a71748-678d-498b-8a4b-6bf26679c1cc-ovn-rundir\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.799306 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3a71748-678d-498b-8a4b-6bf26679c1cc-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.799353 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b3a71748-678d-498b-8a4b-6bf26679c1cc-ovs-rundir\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.799400 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdvfq\" (UniqueName: \"kubernetes.io/projected/b3a71748-678d-498b-8a4b-6bf26679c1cc-kube-api-access-zdvfq\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.799421 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3a71748-678d-498b-8a4b-6bf26679c1cc-combined-ca-bundle\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.799482 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3a71748-678d-498b-8a4b-6bf26679c1cc-config\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.799624 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b3a71748-678d-498b-8a4b-6bf26679c1cc-ovs-rundir\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.799714 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b3a71748-678d-498b-8a4b-6bf26679c1cc-ovn-rundir\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.800829 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3a71748-678d-498b-8a4b-6bf26679c1cc-config\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.803596 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3a71748-678d-498b-8a4b-6bf26679c1cc-combined-ca-bundle\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.806652 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3a71748-678d-498b-8a4b-6bf26679c1cc-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.818100 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdvfq\" (UniqueName: \"kubernetes.io/projected/b3a71748-678d-498b-8a4b-6bf26679c1cc-kube-api-access-zdvfq\") pod \"ovn-controller-metrics-hwlpr\" (UID: \"b3a71748-678d-498b-8a4b-6bf26679c1cc\") " pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.844153 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-h9lfc"] Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.844863 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.872792 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-hwlpr" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.886521 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-q7zqv"] Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.888392 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.892561 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 27 16:42:15 crc kubenswrapper[4814]: I0227 16:42:15.899126 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-q7zqv"] Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.003914 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-config\") pod \"dnsmasq-dns-86db49b7ff-q7zqv\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.003986 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-q7zqv\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.004021 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jqbs\" (UniqueName: \"kubernetes.io/projected/ceb98de6-6418-4892-a542-a4e04080bd24-kube-api-access-9jqbs\") pod \"dnsmasq-dns-86db49b7ff-q7zqv\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.004080 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-q7zqv\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.004209 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-q7zqv\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.105474 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-q7zqv\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.105572 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-q7zqv\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.105618 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-config\") pod \"dnsmasq-dns-86db49b7ff-q7zqv\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.105652 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-q7zqv\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.105695 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jqbs\" (UniqueName: \"kubernetes.io/projected/ceb98de6-6418-4892-a542-a4e04080bd24-kube-api-access-9jqbs\") pod \"dnsmasq-dns-86db49b7ff-q7zqv\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.107715 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-q7zqv\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.107904 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-config\") pod \"dnsmasq-dns-86db49b7ff-q7zqv\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.108361 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-q7zqv\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.108396 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-q7zqv\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.121691 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ea6bae63-9ffa-4ba7-a95f-e686e8d12e87","Type":"ContainerStarted","Data":"0ac79139aa7ded147fb8f300ae7dc00ce9ca99b7042c9497b69766fdb5585b15"} Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.123730 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jqbs\" (UniqueName: \"kubernetes.io/projected/ceb98de6-6418-4892-a542-a4e04080bd24-kube-api-access-9jqbs\") pod \"dnsmasq-dns-86db49b7ff-q7zqv\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.142478 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=20.646294217 podStartE2EDuration="29.142458212s" podCreationTimestamp="2026-02-27 16:41:47 +0000 UTC" firstStartedPulling="2026-02-27 16:41:59.224951646 +0000 UTC m=+1131.677576476" lastFinishedPulling="2026-02-27 16:42:07.721115631 +0000 UTC m=+1140.173740471" observedRunningTime="2026-02-27 16:42:16.140212514 +0000 UTC m=+1148.592837344" watchObservedRunningTime="2026-02-27 16:42:16.142458212 +0000 UTC m=+1148.595083042" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.218807 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.253076 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.275226 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.301476 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-hwlpr"] Feb 27 16:42:16 crc kubenswrapper[4814]: W0227 16:42:16.311225 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3a71748_678d_498b_8a4b_6bf26679c1cc.slice/crio-b70ed4f1de78346c53577f9fe8ead39e9d0333614e8f0b703c1e43eaf6e83677 WatchSource:0}: Error finding container b70ed4f1de78346c53577f9fe8ead39e9d0333614e8f0b703c1e43eaf6e83677: Status 404 returned error can't find the container with id b70ed4f1de78346c53577f9fe8ead39e9d0333614e8f0b703c1e43eaf6e83677 Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.381983 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-h9lfc"] Feb 27 16:42:16 crc kubenswrapper[4814]: I0227 16:42:16.732419 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-q7zqv"] Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.132207 4814 generic.go:334] "Generic (PLEG): container finished" podID="a975048b-7b28-4784-8ae2-cadacba59fdc" containerID="53bdc800d73b94e462252ca71ba59d23382efb8326a165fddcd0023b8503d7ec" exitCode=0 Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.132261 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" event={"ID":"a975048b-7b28-4784-8ae2-cadacba59fdc","Type":"ContainerDied","Data":"53bdc800d73b94e462252ca71ba59d23382efb8326a165fddcd0023b8503d7ec"} Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.132702 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" event={"ID":"a975048b-7b28-4784-8ae2-cadacba59fdc","Type":"ContainerStarted","Data":"934b8833474d1a98c66fc3296b131bfd2923af95b6a924c75349b14dab1e52d7"} Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.136481 4814 generic.go:334] "Generic (PLEG): container finished" podID="ceb98de6-6418-4892-a542-a4e04080bd24" containerID="b3f916f6b9326122c75cf1bc94d79c07fc4b1b4bcb6242245bc9488532c26407" exitCode=0 Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.136607 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" event={"ID":"ceb98de6-6418-4892-a542-a4e04080bd24","Type":"ContainerDied","Data":"b3f916f6b9326122c75cf1bc94d79c07fc4b1b4bcb6242245bc9488532c26407"} Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.136675 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" event={"ID":"ceb98de6-6418-4892-a542-a4e04080bd24","Type":"ContainerStarted","Data":"8c8cf9690df6749185ed5e5721de12deac14b9a6ebc2caef77da7fbb912d4619"} Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.142402 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-hwlpr" event={"ID":"b3a71748-678d-498b-8a4b-6bf26679c1cc","Type":"ContainerStarted","Data":"1060c6e6871db92f6b655baa6fdc80dce9f7904fe784636cfda58babf1bb2ddf"} Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.142456 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-hwlpr" event={"ID":"b3a71748-678d-498b-8a4b-6bf26679c1cc","Type":"ContainerStarted","Data":"b70ed4f1de78346c53577f9fe8ead39e9d0333614e8f0b703c1e43eaf6e83677"} Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.143698 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.217990 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-hwlpr" podStartSLOduration=2.217973753 podStartE2EDuration="2.217973753s" podCreationTimestamp="2026-02-27 16:42:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:42:17.215729674 +0000 UTC m=+1149.668354524" watchObservedRunningTime="2026-02-27 16:42:17.217973753 +0000 UTC m=+1149.670598583" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.227968 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.371184 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.372530 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.476923 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.478857 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.483880 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.488078 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-crdw9" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.488117 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.488160 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.488731 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.505881 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.563768 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-dns-svc\") pod \"a975048b-7b28-4784-8ae2-cadacba59fdc\" (UID: \"a975048b-7b28-4784-8ae2-cadacba59fdc\") " Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.563810 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-config\") pod \"a975048b-7b28-4784-8ae2-cadacba59fdc\" (UID: \"a975048b-7b28-4784-8ae2-cadacba59fdc\") " Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.563858 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lv94f\" (UniqueName: \"kubernetes.io/projected/a975048b-7b28-4784-8ae2-cadacba59fdc-kube-api-access-lv94f\") pod \"a975048b-7b28-4784-8ae2-cadacba59fdc\" (UID: \"a975048b-7b28-4784-8ae2-cadacba59fdc\") " Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.563889 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-ovsdbserver-sb\") pod \"a975048b-7b28-4784-8ae2-cadacba59fdc\" (UID: \"a975048b-7b28-4784-8ae2-cadacba59fdc\") " Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.564136 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/381df711-c6a7-428a-94d5-8f2f84b51ef6-scripts\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.564162 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/381df711-c6a7-428a-94d5-8f2f84b51ef6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.564203 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/381df711-c6a7-428a-94d5-8f2f84b51ef6-config\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.564217 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/381df711-c6a7-428a-94d5-8f2f84b51ef6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.564239 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/381df711-c6a7-428a-94d5-8f2f84b51ef6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.564297 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xclpn\" (UniqueName: \"kubernetes.io/projected/381df711-c6a7-428a-94d5-8f2f84b51ef6-kube-api-access-xclpn\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.564320 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/381df711-c6a7-428a-94d5-8f2f84b51ef6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.568452 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a975048b-7b28-4784-8ae2-cadacba59fdc-kube-api-access-lv94f" (OuterVolumeSpecName: "kube-api-access-lv94f") pod "a975048b-7b28-4784-8ae2-cadacba59fdc" (UID: "a975048b-7b28-4784-8ae2-cadacba59fdc"). InnerVolumeSpecName "kube-api-access-lv94f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.580956 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-config" (OuterVolumeSpecName: "config") pod "a975048b-7b28-4784-8ae2-cadacba59fdc" (UID: "a975048b-7b28-4784-8ae2-cadacba59fdc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.581001 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a975048b-7b28-4784-8ae2-cadacba59fdc" (UID: "a975048b-7b28-4784-8ae2-cadacba59fdc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.585823 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a975048b-7b28-4784-8ae2-cadacba59fdc" (UID: "a975048b-7b28-4784-8ae2-cadacba59fdc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.665371 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xclpn\" (UniqueName: \"kubernetes.io/projected/381df711-c6a7-428a-94d5-8f2f84b51ef6-kube-api-access-xclpn\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.665422 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/381df711-c6a7-428a-94d5-8f2f84b51ef6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.665475 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/381df711-c6a7-428a-94d5-8f2f84b51ef6-scripts\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.665495 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/381df711-c6a7-428a-94d5-8f2f84b51ef6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.665532 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/381df711-c6a7-428a-94d5-8f2f84b51ef6-config\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.665547 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/381df711-c6a7-428a-94d5-8f2f84b51ef6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.665566 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/381df711-c6a7-428a-94d5-8f2f84b51ef6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.665633 4814 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.665643 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.665653 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lv94f\" (UniqueName: \"kubernetes.io/projected/a975048b-7b28-4784-8ae2-cadacba59fdc-kube-api-access-lv94f\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.665661 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a975048b-7b28-4784-8ae2-cadacba59fdc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.667003 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/381df711-c6a7-428a-94d5-8f2f84b51ef6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.667433 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/381df711-c6a7-428a-94d5-8f2f84b51ef6-scripts\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.668108 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/381df711-c6a7-428a-94d5-8f2f84b51ef6-config\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.669984 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/381df711-c6a7-428a-94d5-8f2f84b51ef6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.670588 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/381df711-c6a7-428a-94d5-8f2f84b51ef6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.671325 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/381df711-c6a7-428a-94d5-8f2f84b51ef6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.691223 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xclpn\" (UniqueName: \"kubernetes.io/projected/381df711-c6a7-428a-94d5-8f2f84b51ef6-kube-api-access-xclpn\") pod \"ovn-northd-0\" (UID: \"381df711-c6a7-428a-94d5-8f2f84b51ef6\") " pod="openstack/ovn-northd-0" Feb 27 16:42:17 crc kubenswrapper[4814]: I0227 16:42:17.831068 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 27 16:42:18 crc kubenswrapper[4814]: I0227 16:42:18.156628 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" event={"ID":"a975048b-7b28-4784-8ae2-cadacba59fdc","Type":"ContainerDied","Data":"934b8833474d1a98c66fc3296b131bfd2923af95b6a924c75349b14dab1e52d7"} Feb 27 16:42:18 crc kubenswrapper[4814]: I0227 16:42:18.156925 4814 scope.go:117] "RemoveContainer" containerID="53bdc800d73b94e462252ca71ba59d23382efb8326a165fddcd0023b8503d7ec" Feb 27 16:42:18 crc kubenswrapper[4814]: I0227 16:42:18.156668 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-h9lfc" Feb 27 16:42:18 crc kubenswrapper[4814]: I0227 16:42:18.158911 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" event={"ID":"ceb98de6-6418-4892-a542-a4e04080bd24","Type":"ContainerStarted","Data":"298865db040dc9e5450c5a2766d3d6f40e0ffdf935e646a01e543bba41eed1dd"} Feb 27 16:42:18 crc kubenswrapper[4814]: I0227 16:42:18.187621 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" podStartSLOduration=3.187601356 podStartE2EDuration="3.187601356s" podCreationTimestamp="2026-02-27 16:42:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:42:18.18016314 +0000 UTC m=+1150.632787990" watchObservedRunningTime="2026-02-27 16:42:18.187601356 +0000 UTC m=+1150.640226186" Feb 27 16:42:18 crc kubenswrapper[4814]: I0227 16:42:18.245974 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-h9lfc"] Feb 27 16:42:18 crc kubenswrapper[4814]: I0227 16:42:18.253041 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-h9lfc"] Feb 27 16:42:18 crc kubenswrapper[4814]: I0227 16:42:18.299830 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 27 16:42:18 crc kubenswrapper[4814]: I0227 16:42:18.308992 4814 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 16:42:18 crc kubenswrapper[4814]: I0227 16:42:18.509392 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a975048b-7b28-4784-8ae2-cadacba59fdc" path="/var/lib/kubelet/pods/a975048b-7b28-4784-8ae2-cadacba59fdc/volumes" Feb 27 16:42:18 crc kubenswrapper[4814]: I0227 16:42:18.688849 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 27 16:42:18 crc kubenswrapper[4814]: I0227 16:42:18.688910 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 27 16:42:19 crc kubenswrapper[4814]: I0227 16:42:19.000743 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 27 16:42:19 crc kubenswrapper[4814]: I0227 16:42:19.167218 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"381df711-c6a7-428a-94d5-8f2f84b51ef6","Type":"ContainerStarted","Data":"cf23d2b5b222ade817d5821eac477c15d3de50d68c2d77d471ff2001f3ca4c36"} Feb 27 16:42:19 crc kubenswrapper[4814]: I0227 16:42:19.167734 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.399938 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.748645 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-q7zqv"] Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.750364 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" podUID="ceb98de6-6418-4892-a542-a4e04080bd24" containerName="dnsmasq-dns" containerID="cri-o://298865db040dc9e5450c5a2766d3d6f40e0ffdf935e646a01e543bba41eed1dd" gracePeriod=10 Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.777425 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-p6sq9"] Feb 27 16:42:21 crc kubenswrapper[4814]: E0227 16:42:21.777689 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a975048b-7b28-4784-8ae2-cadacba59fdc" containerName="init" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.777700 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="a975048b-7b28-4784-8ae2-cadacba59fdc" containerName="init" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.777857 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="a975048b-7b28-4784-8ae2-cadacba59fdc" containerName="init" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.778569 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.829454 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-p6sq9\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.829740 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-p6sq9\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.829820 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-dns-svc\") pod \"dnsmasq-dns-698758b865-p6sq9\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.829908 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx4vj\" (UniqueName: \"kubernetes.io/projected/958486c1-15d1-4184-9197-84fd9c0ba31b-kube-api-access-hx4vj\") pod \"dnsmasq-dns-698758b865-p6sq9\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.830001 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-config\") pod \"dnsmasq-dns-698758b865-p6sq9\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.844062 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-p6sq9"] Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.931314 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-p6sq9\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.931556 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-dns-svc\") pod \"dnsmasq-dns-698758b865-p6sq9\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.931688 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx4vj\" (UniqueName: \"kubernetes.io/projected/958486c1-15d1-4184-9197-84fd9c0ba31b-kube-api-access-hx4vj\") pod \"dnsmasq-dns-698758b865-p6sq9\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.931799 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-config\") pod \"dnsmasq-dns-698758b865-p6sq9\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.931920 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-p6sq9\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.932729 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-p6sq9\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.932749 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-dns-svc\") pod \"dnsmasq-dns-698758b865-p6sq9\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.933287 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-config\") pod \"dnsmasq-dns-698758b865-p6sq9\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.933509 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-p6sq9\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:21 crc kubenswrapper[4814]: I0227 16:42:21.955807 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx4vj\" (UniqueName: \"kubernetes.io/projected/958486c1-15d1-4184-9197-84fd9c0ba31b-kube-api-access-hx4vj\") pod \"dnsmasq-dns-698758b865-p6sq9\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:22 crc kubenswrapper[4814]: I0227 16:42:22.102074 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:22 crc kubenswrapper[4814]: I0227 16:42:22.197814 4814 generic.go:334] "Generic (PLEG): container finished" podID="ceb98de6-6418-4892-a542-a4e04080bd24" containerID="298865db040dc9e5450c5a2766d3d6f40e0ffdf935e646a01e543bba41eed1dd" exitCode=0 Feb 27 16:42:22 crc kubenswrapper[4814]: I0227 16:42:22.197976 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" event={"ID":"ceb98de6-6418-4892-a542-a4e04080bd24","Type":"ContainerDied","Data":"298865db040dc9e5450c5a2766d3d6f40e0ffdf935e646a01e543bba41eed1dd"} Feb 27 16:42:22 crc kubenswrapper[4814]: I0227 16:42:22.646346 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-p6sq9"] Feb 27 16:42:22 crc kubenswrapper[4814]: I0227 16:42:22.902923 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:42:22 crc kubenswrapper[4814]: I0227 16:42:22.902986 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:42:22 crc kubenswrapper[4814]: I0227 16:42:22.903036 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:42:22 crc kubenswrapper[4814]: I0227 16:42:22.904021 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"be2cf4df4208edbaae217431871265acc14c77fe4583a6ab07012775c0d5e5f8"} pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 16:42:22 crc kubenswrapper[4814]: I0227 16:42:22.904082 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" containerID="cri-o://be2cf4df4208edbaae217431871265acc14c77fe4583a6ab07012775c0d5e5f8" gracePeriod=600 Feb 27 16:42:22 crc kubenswrapper[4814]: I0227 16:42:22.917802 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 27 16:42:22 crc kubenswrapper[4814]: I0227 16:42:22.926989 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 27 16:42:22 crc kubenswrapper[4814]: I0227 16:42:22.941588 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 27 16:42:22 crc kubenswrapper[4814]: I0227 16:42:22.942276 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-8fcx8" Feb 27 16:42:22 crc kubenswrapper[4814]: I0227 16:42:22.942458 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 27 16:42:22 crc kubenswrapper[4814]: I0227 16:42:22.942698 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 27 16:42:22 crc kubenswrapper[4814]: I0227 16:42:22.969676 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.058623 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpnsb\" (UniqueName: \"kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-kube-api-access-tpnsb\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.058724 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.058756 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.058780 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9746d8ab-e2c8-419c-8327-6705260b31dd-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.058799 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9746d8ab-e2c8-419c-8327-6705260b31dd-cache\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.058835 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9746d8ab-e2c8-419c-8327-6705260b31dd-lock\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.159880 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.159931 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.159955 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9746d8ab-e2c8-419c-8327-6705260b31dd-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.159974 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9746d8ab-e2c8-419c-8327-6705260b31dd-cache\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.160005 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9746d8ab-e2c8-419c-8327-6705260b31dd-lock\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.160059 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpnsb\" (UniqueName: \"kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-kube-api-access-tpnsb\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: E0227 16:42:23.160453 4814 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 27 16:42:23 crc kubenswrapper[4814]: E0227 16:42:23.160471 4814 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 27 16:42:23 crc kubenswrapper[4814]: E0227 16:42:23.160506 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift podName:9746d8ab-e2c8-419c-8327-6705260b31dd nodeName:}" failed. No retries permitted until 2026-02-27 16:42:23.660492392 +0000 UTC m=+1156.113117222 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift") pod "swift-storage-0" (UID: "9746d8ab-e2c8-419c-8327-6705260b31dd") : configmap "swift-ring-files" not found Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.160844 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.162372 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9746d8ab-e2c8-419c-8327-6705260b31dd-lock\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.162715 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9746d8ab-e2c8-419c-8327-6705260b31dd-cache\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.171106 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9746d8ab-e2c8-419c-8327-6705260b31dd-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.193791 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpnsb\" (UniqueName: \"kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-kube-api-access-tpnsb\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.204853 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-p6sq9" event={"ID":"958486c1-15d1-4184-9197-84fd9c0ba31b","Type":"ContainerStarted","Data":"abb6015521fefb10a106eea8fc30e0578835eca45614bb6b121b64214432cc2a"} Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.211156 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.314002 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.464941 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-ovsdbserver-sb\") pod \"ceb98de6-6418-4892-a542-a4e04080bd24\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.464986 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-config\") pod \"ceb98de6-6418-4892-a542-a4e04080bd24\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.465019 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-dns-svc\") pod \"ceb98de6-6418-4892-a542-a4e04080bd24\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.465061 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jqbs\" (UniqueName: \"kubernetes.io/projected/ceb98de6-6418-4892-a542-a4e04080bd24-kube-api-access-9jqbs\") pod \"ceb98de6-6418-4892-a542-a4e04080bd24\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.465174 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-ovsdbserver-nb\") pod \"ceb98de6-6418-4892-a542-a4e04080bd24\" (UID: \"ceb98de6-6418-4892-a542-a4e04080bd24\") " Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.468925 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceb98de6-6418-4892-a542-a4e04080bd24-kube-api-access-9jqbs" (OuterVolumeSpecName: "kube-api-access-9jqbs") pod "ceb98de6-6418-4892-a542-a4e04080bd24" (UID: "ceb98de6-6418-4892-a542-a4e04080bd24"). InnerVolumeSpecName "kube-api-access-9jqbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.505278 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ceb98de6-6418-4892-a542-a4e04080bd24" (UID: "ceb98de6-6418-4892-a542-a4e04080bd24"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.506238 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-config" (OuterVolumeSpecName: "config") pod "ceb98de6-6418-4892-a542-a4e04080bd24" (UID: "ceb98de6-6418-4892-a542-a4e04080bd24"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.513926 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ceb98de6-6418-4892-a542-a4e04080bd24" (UID: "ceb98de6-6418-4892-a542-a4e04080bd24"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.514338 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ceb98de6-6418-4892-a542-a4e04080bd24" (UID: "ceb98de6-6418-4892-a542-a4e04080bd24"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.566714 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.566745 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.566753 4814 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.566762 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jqbs\" (UniqueName: \"kubernetes.io/projected/ceb98de6-6418-4892-a542-a4e04080bd24-kube-api-access-9jqbs\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.566773 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ceb98de6-6418-4892-a542-a4e04080bd24-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.655167 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.668208 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:23 crc kubenswrapper[4814]: E0227 16:42:23.668416 4814 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 27 16:42:23 crc kubenswrapper[4814]: E0227 16:42:23.668431 4814 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 27 16:42:23 crc kubenswrapper[4814]: E0227 16:42:23.668481 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift podName:9746d8ab-e2c8-419c-8327-6705260b31dd nodeName:}" failed. No retries permitted until 2026-02-27 16:42:24.668464987 +0000 UTC m=+1157.121089817 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift") pod "swift-storage-0" (UID: "9746d8ab-e2c8-419c-8327-6705260b31dd") : configmap "swift-ring-files" not found Feb 27 16:42:23 crc kubenswrapper[4814]: I0227 16:42:23.758619 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 27 16:42:24 crc kubenswrapper[4814]: I0227 16:42:24.214619 4814 generic.go:334] "Generic (PLEG): container finished" podID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerID="be2cf4df4208edbaae217431871265acc14c77fe4583a6ab07012775c0d5e5f8" exitCode=0 Feb 27 16:42:24 crc kubenswrapper[4814]: I0227 16:42:24.214705 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerDied","Data":"be2cf4df4208edbaae217431871265acc14c77fe4583a6ab07012775c0d5e5f8"} Feb 27 16:42:24 crc kubenswrapper[4814]: I0227 16:42:24.215022 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"5e8afb4996159ac03d9ce86a713b7b4aae36d378d653cf752342458cdea88224"} Feb 27 16:42:24 crc kubenswrapper[4814]: I0227 16:42:24.215045 4814 scope.go:117] "RemoveContainer" containerID="7a221694cd83011f5779a509b7774635f372c6adcfadbeae289b463611af2c58" Feb 27 16:42:24 crc kubenswrapper[4814]: I0227 16:42:24.216657 4814 generic.go:334] "Generic (PLEG): container finished" podID="958486c1-15d1-4184-9197-84fd9c0ba31b" containerID="749ff4565088119df234b31211740010acefc4e5ed9665a0e594cb14b4cf5724" exitCode=0 Feb 27 16:42:24 crc kubenswrapper[4814]: I0227 16:42:24.216707 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-p6sq9" event={"ID":"958486c1-15d1-4184-9197-84fd9c0ba31b","Type":"ContainerDied","Data":"749ff4565088119df234b31211740010acefc4e5ed9665a0e594cb14b4cf5724"} Feb 27 16:42:24 crc kubenswrapper[4814]: I0227 16:42:24.220483 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" Feb 27 16:42:24 crc kubenswrapper[4814]: I0227 16:42:24.223390 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-q7zqv" event={"ID":"ceb98de6-6418-4892-a542-a4e04080bd24","Type":"ContainerDied","Data":"8c8cf9690df6749185ed5e5721de12deac14b9a6ebc2caef77da7fbb912d4619"} Feb 27 16:42:24 crc kubenswrapper[4814]: I0227 16:42:24.248061 4814 scope.go:117] "RemoveContainer" containerID="298865db040dc9e5450c5a2766d3d6f40e0ffdf935e646a01e543bba41eed1dd" Feb 27 16:42:24 crc kubenswrapper[4814]: I0227 16:42:24.294556 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-q7zqv"] Feb 27 16:42:24 crc kubenswrapper[4814]: I0227 16:42:24.295972 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-q7zqv"] Feb 27 16:42:24 crc kubenswrapper[4814]: I0227 16:42:24.318439 4814 scope.go:117] "RemoveContainer" containerID="b3f916f6b9326122c75cf1bc94d79c07fc4b1b4bcb6242245bc9488532c26407" Feb 27 16:42:24 crc kubenswrapper[4814]: I0227 16:42:24.496002 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ceb98de6-6418-4892-a542-a4e04080bd24" path="/var/lib/kubelet/pods/ceb98de6-6418-4892-a542-a4e04080bd24/volumes" Feb 27 16:42:24 crc kubenswrapper[4814]: I0227 16:42:24.686970 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:24 crc kubenswrapper[4814]: E0227 16:42:24.687156 4814 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 27 16:42:24 crc kubenswrapper[4814]: E0227 16:42:24.687182 4814 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 27 16:42:24 crc kubenswrapper[4814]: E0227 16:42:24.687231 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift podName:9746d8ab-e2c8-419c-8327-6705260b31dd nodeName:}" failed. No retries permitted until 2026-02-27 16:42:26.687214293 +0000 UTC m=+1159.139839113 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift") pod "swift-storage-0" (UID: "9746d8ab-e2c8-419c-8327-6705260b31dd") : configmap "swift-ring-files" not found Feb 27 16:42:25 crc kubenswrapper[4814]: I0227 16:42:25.233555 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-p6sq9" event={"ID":"958486c1-15d1-4184-9197-84fd9c0ba31b","Type":"ContainerStarted","Data":"6669b50d32f8a4e6280bb46209b22c07a78970e4deecb6329c65ddeeaea0a094"} Feb 27 16:42:25 crc kubenswrapper[4814]: I0227 16:42:25.235164 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:25 crc kubenswrapper[4814]: I0227 16:42:25.242429 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"381df711-c6a7-428a-94d5-8f2f84b51ef6","Type":"ContainerStarted","Data":"5ef5296c15e3e285f45c8fc7a1ce6619ca42d03a2f83ce3c3c9fe54d1ace167a"} Feb 27 16:42:25 crc kubenswrapper[4814]: I0227 16:42:25.242494 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"381df711-c6a7-428a-94d5-8f2f84b51ef6","Type":"ContainerStarted","Data":"54611092f4f35415262e42be79037287e58f7d2b10bcd489f6cd918d46a09cc0"} Feb 27 16:42:25 crc kubenswrapper[4814]: I0227 16:42:25.243185 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 27 16:42:25 crc kubenswrapper[4814]: I0227 16:42:25.266240 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-p6sq9" podStartSLOduration=4.266217516 podStartE2EDuration="4.266217516s" podCreationTimestamp="2026-02-27 16:42:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:42:25.25385832 +0000 UTC m=+1157.706483190" watchObservedRunningTime="2026-02-27 16:42:25.266217516 +0000 UTC m=+1157.718842376" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.156138 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.463962656 podStartE2EDuration="9.156112297s" podCreationTimestamp="2026-02-27 16:42:17 +0000 UTC" firstStartedPulling="2026-02-27 16:42:18.308810829 +0000 UTC m=+1150.761435659" lastFinishedPulling="2026-02-27 16:42:24.00096046 +0000 UTC m=+1156.453585300" observedRunningTime="2026-02-27 16:42:25.286194744 +0000 UTC m=+1157.738819604" watchObservedRunningTime="2026-02-27 16:42:26.156112297 +0000 UTC m=+1158.608737157" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.156836 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-64zxx"] Feb 27 16:42:26 crc kubenswrapper[4814]: E0227 16:42:26.157490 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb98de6-6418-4892-a542-a4e04080bd24" containerName="init" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.157521 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb98de6-6418-4892-a542-a4e04080bd24" containerName="init" Feb 27 16:42:26 crc kubenswrapper[4814]: E0227 16:42:26.157591 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb98de6-6418-4892-a542-a4e04080bd24" containerName="dnsmasq-dns" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.157613 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb98de6-6418-4892-a542-a4e04080bd24" containerName="dnsmasq-dns" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.158026 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceb98de6-6418-4892-a542-a4e04080bd24" containerName="dnsmasq-dns" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.158951 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-64zxx" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.164154 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.167116 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-64zxx"] Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.317054 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78hr6\" (UniqueName: \"kubernetes.io/projected/110fcb3a-31c0-47d9-9e42-7e534e5ede02-kube-api-access-78hr6\") pod \"root-account-create-update-64zxx\" (UID: \"110fcb3a-31c0-47d9-9e42-7e534e5ede02\") " pod="openstack/root-account-create-update-64zxx" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.317141 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/110fcb3a-31c0-47d9-9e42-7e534e5ede02-operator-scripts\") pod \"root-account-create-update-64zxx\" (UID: \"110fcb3a-31c0-47d9-9e42-7e534e5ede02\") " pod="openstack/root-account-create-update-64zxx" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.419462 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78hr6\" (UniqueName: \"kubernetes.io/projected/110fcb3a-31c0-47d9-9e42-7e534e5ede02-kube-api-access-78hr6\") pod \"root-account-create-update-64zxx\" (UID: \"110fcb3a-31c0-47d9-9e42-7e534e5ede02\") " pod="openstack/root-account-create-update-64zxx" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.419537 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/110fcb3a-31c0-47d9-9e42-7e534e5ede02-operator-scripts\") pod \"root-account-create-update-64zxx\" (UID: \"110fcb3a-31c0-47d9-9e42-7e534e5ede02\") " pod="openstack/root-account-create-update-64zxx" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.421694 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/110fcb3a-31c0-47d9-9e42-7e534e5ede02-operator-scripts\") pod \"root-account-create-update-64zxx\" (UID: \"110fcb3a-31c0-47d9-9e42-7e534e5ede02\") " pod="openstack/root-account-create-update-64zxx" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.439502 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78hr6\" (UniqueName: \"kubernetes.io/projected/110fcb3a-31c0-47d9-9e42-7e534e5ede02-kube-api-access-78hr6\") pod \"root-account-create-update-64zxx\" (UID: \"110fcb3a-31c0-47d9-9e42-7e534e5ede02\") " pod="openstack/root-account-create-update-64zxx" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.484318 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-64zxx" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.523509 4814 scope.go:117] "RemoveContainer" containerID="164ebc8641f4885df5b56586f8ce5d02d3bd9dbba699838e0448b5fbb8d90f12" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.724080 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:26 crc kubenswrapper[4814]: E0227 16:42:26.724853 4814 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 27 16:42:26 crc kubenswrapper[4814]: E0227 16:42:26.724872 4814 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 27 16:42:26 crc kubenswrapper[4814]: E0227 16:42:26.724932 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift podName:9746d8ab-e2c8-419c-8327-6705260b31dd nodeName:}" failed. No retries permitted until 2026-02-27 16:42:30.72491478 +0000 UTC m=+1163.177539610 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift") pod "swift-storage-0" (UID: "9746d8ab-e2c8-419c-8327-6705260b31dd") : configmap "swift-ring-files" not found Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.825968 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.872875 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-k6ccx"] Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.874838 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.877776 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.877886 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.878072 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.888282 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-k6ccx"] Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.925869 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 27 16:42:26 crc kubenswrapper[4814]: W0227 16:42:26.957925 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod110fcb3a_31c0_47d9_9e42_7e534e5ede02.slice/crio-606c87698face0dead555133dafc8298d58f257c2d4b57407b02386aa764c43d WatchSource:0}: Error finding container 606c87698face0dead555133dafc8298d58f257c2d4b57407b02386aa764c43d: Status 404 returned error can't find the container with id 606c87698face0dead555133dafc8298d58f257c2d4b57407b02386aa764c43d Feb 27 16:42:26 crc kubenswrapper[4814]: I0227 16:42:26.971063 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-64zxx"] Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.032113 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-swiftconf\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.032171 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3daf276e-d557-4192-831c-d2aa124fe9bc-etc-swift\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.032197 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-combined-ca-bundle\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.032219 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsnl2\" (UniqueName: \"kubernetes.io/projected/3daf276e-d557-4192-831c-d2aa124fe9bc-kube-api-access-tsnl2\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.032280 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3daf276e-d557-4192-831c-d2aa124fe9bc-scripts\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.032349 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3daf276e-d557-4192-831c-d2aa124fe9bc-ring-data-devices\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.032633 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-dispersionconf\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.134336 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-swiftconf\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.134402 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3daf276e-d557-4192-831c-d2aa124fe9bc-etc-swift\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.134436 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-combined-ca-bundle\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.134464 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsnl2\" (UniqueName: \"kubernetes.io/projected/3daf276e-d557-4192-831c-d2aa124fe9bc-kube-api-access-tsnl2\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.134491 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3daf276e-d557-4192-831c-d2aa124fe9bc-scripts\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.134510 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3daf276e-d557-4192-831c-d2aa124fe9bc-ring-data-devices\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.134589 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-dispersionconf\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.134780 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3daf276e-d557-4192-831c-d2aa124fe9bc-etc-swift\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.135417 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3daf276e-d557-4192-831c-d2aa124fe9bc-scripts\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.135801 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3daf276e-d557-4192-831c-d2aa124fe9bc-ring-data-devices\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.139522 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-dispersionconf\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.139997 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-combined-ca-bundle\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.140809 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-swiftconf\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.151901 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsnl2\" (UniqueName: \"kubernetes.io/projected/3daf276e-d557-4192-831c-d2aa124fe9bc-kube-api-access-tsnl2\") pod \"swift-ring-rebalance-k6ccx\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.196913 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.268068 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-64zxx" event={"ID":"110fcb3a-31c0-47d9-9e42-7e534e5ede02","Type":"ContainerStarted","Data":"224f2879b4b05aa596bc464639c7088cf0190a24804cdf9cadaff1d692aae3b3"} Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.268114 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-64zxx" event={"ID":"110fcb3a-31c0-47d9-9e42-7e534e5ede02","Type":"ContainerStarted","Data":"606c87698face0dead555133dafc8298d58f257c2d4b57407b02386aa764c43d"} Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.302233 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-64zxx" podStartSLOduration=1.302214913 podStartE2EDuration="1.302214913s" podCreationTimestamp="2026-02-27 16:42:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:42:27.297000483 +0000 UTC m=+1159.749625313" watchObservedRunningTime="2026-02-27 16:42:27.302214913 +0000 UTC m=+1159.754839743" Feb 27 16:42:27 crc kubenswrapper[4814]: I0227 16:42:27.699352 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-k6ccx"] Feb 27 16:42:27 crc kubenswrapper[4814]: W0227 16:42:27.701536 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3daf276e_d557_4192_831c_d2aa124fe9bc.slice/crio-8dc718481d252c2a39c32c2c35abea68932b1953547dce0055be1518a92d15ba WatchSource:0}: Error finding container 8dc718481d252c2a39c32c2c35abea68932b1953547dce0055be1518a92d15ba: Status 404 returned error can't find the container with id 8dc718481d252c2a39c32c2c35abea68932b1953547dce0055be1518a92d15ba Feb 27 16:42:28 crc kubenswrapper[4814]: I0227 16:42:28.278093 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-k6ccx" event={"ID":"3daf276e-d557-4192-831c-d2aa124fe9bc","Type":"ContainerStarted","Data":"8dc718481d252c2a39c32c2c35abea68932b1953547dce0055be1518a92d15ba"} Feb 27 16:42:28 crc kubenswrapper[4814]: I0227 16:42:28.281005 4814 generic.go:334] "Generic (PLEG): container finished" podID="110fcb3a-31c0-47d9-9e42-7e534e5ede02" containerID="224f2879b4b05aa596bc464639c7088cf0190a24804cdf9cadaff1d692aae3b3" exitCode=0 Feb 27 16:42:28 crc kubenswrapper[4814]: I0227 16:42:28.281046 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-64zxx" event={"ID":"110fcb3a-31c0-47d9-9e42-7e534e5ede02","Type":"ContainerDied","Data":"224f2879b4b05aa596bc464639c7088cf0190a24804cdf9cadaff1d692aae3b3"} Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.136508 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-8tfhv"] Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.137627 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8tfhv" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.145898 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8tfhv"] Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.260065 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-9481-account-create-update-rzblc"] Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.261592 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9481-account-create-update-rzblc" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.263376 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.268315 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a048b31-2392-4739-8350-93136b0820b3-operator-scripts\") pod \"glance-db-create-8tfhv\" (UID: \"7a048b31-2392-4739-8350-93136b0820b3\") " pod="openstack/glance-db-create-8tfhv" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.268377 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skrcs\" (UniqueName: \"kubernetes.io/projected/7a048b31-2392-4739-8350-93136b0820b3-kube-api-access-skrcs\") pod \"glance-db-create-8tfhv\" (UID: \"7a048b31-2392-4739-8350-93136b0820b3\") " pod="openstack/glance-db-create-8tfhv" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.270619 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-9481-account-create-update-rzblc"] Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.369885 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skrcs\" (UniqueName: \"kubernetes.io/projected/7a048b31-2392-4739-8350-93136b0820b3-kube-api-access-skrcs\") pod \"glance-db-create-8tfhv\" (UID: \"7a048b31-2392-4739-8350-93136b0820b3\") " pod="openstack/glance-db-create-8tfhv" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.370424 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/653257dd-39a8-4062-b70f-33d78ef04baf-operator-scripts\") pod \"glance-9481-account-create-update-rzblc\" (UID: \"653257dd-39a8-4062-b70f-33d78ef04baf\") " pod="openstack/glance-9481-account-create-update-rzblc" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.370528 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpnv6\" (UniqueName: \"kubernetes.io/projected/653257dd-39a8-4062-b70f-33d78ef04baf-kube-api-access-xpnv6\") pod \"glance-9481-account-create-update-rzblc\" (UID: \"653257dd-39a8-4062-b70f-33d78ef04baf\") " pod="openstack/glance-9481-account-create-update-rzblc" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.370558 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a048b31-2392-4739-8350-93136b0820b3-operator-scripts\") pod \"glance-db-create-8tfhv\" (UID: \"7a048b31-2392-4739-8350-93136b0820b3\") " pod="openstack/glance-db-create-8tfhv" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.371313 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a048b31-2392-4739-8350-93136b0820b3-operator-scripts\") pod \"glance-db-create-8tfhv\" (UID: \"7a048b31-2392-4739-8350-93136b0820b3\") " pod="openstack/glance-db-create-8tfhv" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.395964 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skrcs\" (UniqueName: \"kubernetes.io/projected/7a048b31-2392-4739-8350-93136b0820b3-kube-api-access-skrcs\") pod \"glance-db-create-8tfhv\" (UID: \"7a048b31-2392-4739-8350-93136b0820b3\") " pod="openstack/glance-db-create-8tfhv" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.463388 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8tfhv" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.472172 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/653257dd-39a8-4062-b70f-33d78ef04baf-operator-scripts\") pod \"glance-9481-account-create-update-rzblc\" (UID: \"653257dd-39a8-4062-b70f-33d78ef04baf\") " pod="openstack/glance-9481-account-create-update-rzblc" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.472284 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpnv6\" (UniqueName: \"kubernetes.io/projected/653257dd-39a8-4062-b70f-33d78ef04baf-kube-api-access-xpnv6\") pod \"glance-9481-account-create-update-rzblc\" (UID: \"653257dd-39a8-4062-b70f-33d78ef04baf\") " pod="openstack/glance-9481-account-create-update-rzblc" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.472959 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/653257dd-39a8-4062-b70f-33d78ef04baf-operator-scripts\") pod \"glance-9481-account-create-update-rzblc\" (UID: \"653257dd-39a8-4062-b70f-33d78ef04baf\") " pod="openstack/glance-9481-account-create-update-rzblc" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.491939 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpnv6\" (UniqueName: \"kubernetes.io/projected/653257dd-39a8-4062-b70f-33d78ef04baf-kube-api-access-xpnv6\") pod \"glance-9481-account-create-update-rzblc\" (UID: \"653257dd-39a8-4062-b70f-33d78ef04baf\") " pod="openstack/glance-9481-account-create-update-rzblc" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.584645 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9481-account-create-update-rzblc" Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.992152 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-j5g9f"] Feb 27 16:42:29 crc kubenswrapper[4814]: I0227 16:42:29.993586 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-j5g9f" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.004099 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-j5g9f"] Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.037396 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-64zxx" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.084856 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78hr6\" (UniqueName: \"kubernetes.io/projected/110fcb3a-31c0-47d9-9e42-7e534e5ede02-kube-api-access-78hr6\") pod \"110fcb3a-31c0-47d9-9e42-7e534e5ede02\" (UID: \"110fcb3a-31c0-47d9-9e42-7e534e5ede02\") " Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.084952 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/110fcb3a-31c0-47d9-9e42-7e534e5ede02-operator-scripts\") pod \"110fcb3a-31c0-47d9-9e42-7e534e5ede02\" (UID: \"110fcb3a-31c0-47d9-9e42-7e534e5ede02\") " Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.085542 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k4dv\" (UniqueName: \"kubernetes.io/projected/0c835cab-f5ff-497c-a4c3-865e2a30f48f-kube-api-access-8k4dv\") pod \"keystone-db-create-j5g9f\" (UID: \"0c835cab-f5ff-497c-a4c3-865e2a30f48f\") " pod="openstack/keystone-db-create-j5g9f" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.085668 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c835cab-f5ff-497c-a4c3-865e2a30f48f-operator-scripts\") pod \"keystone-db-create-j5g9f\" (UID: \"0c835cab-f5ff-497c-a4c3-865e2a30f48f\") " pod="openstack/keystone-db-create-j5g9f" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.087894 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/110fcb3a-31c0-47d9-9e42-7e534e5ede02-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "110fcb3a-31c0-47d9-9e42-7e534e5ede02" (UID: "110fcb3a-31c0-47d9-9e42-7e534e5ede02"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.099537 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/110fcb3a-31c0-47d9-9e42-7e534e5ede02-kube-api-access-78hr6" (OuterVolumeSpecName: "kube-api-access-78hr6") pod "110fcb3a-31c0-47d9-9e42-7e534e5ede02" (UID: "110fcb3a-31c0-47d9-9e42-7e534e5ede02"). InnerVolumeSpecName "kube-api-access-78hr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.149310 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-9b47-account-create-update-8bqfg"] Feb 27 16:42:30 crc kubenswrapper[4814]: E0227 16:42:30.149699 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="110fcb3a-31c0-47d9-9e42-7e534e5ede02" containerName="mariadb-account-create-update" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.149715 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="110fcb3a-31c0-47d9-9e42-7e534e5ede02" containerName="mariadb-account-create-update" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.149861 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="110fcb3a-31c0-47d9-9e42-7e534e5ede02" containerName="mariadb-account-create-update" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.150473 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9b47-account-create-update-8bqfg" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.153665 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.157421 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9b47-account-create-update-8bqfg"] Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.190170 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k4dv\" (UniqueName: \"kubernetes.io/projected/0c835cab-f5ff-497c-a4c3-865e2a30f48f-kube-api-access-8k4dv\") pod \"keystone-db-create-j5g9f\" (UID: \"0c835cab-f5ff-497c-a4c3-865e2a30f48f\") " pod="openstack/keystone-db-create-j5g9f" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.190329 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c835cab-f5ff-497c-a4c3-865e2a30f48f-operator-scripts\") pod \"keystone-db-create-j5g9f\" (UID: \"0c835cab-f5ff-497c-a4c3-865e2a30f48f\") " pod="openstack/keystone-db-create-j5g9f" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.190383 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/110fcb3a-31c0-47d9-9e42-7e534e5ede02-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.190396 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78hr6\" (UniqueName: \"kubernetes.io/projected/110fcb3a-31c0-47d9-9e42-7e534e5ede02-kube-api-access-78hr6\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.190977 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c835cab-f5ff-497c-a4c3-865e2a30f48f-operator-scripts\") pod \"keystone-db-create-j5g9f\" (UID: \"0c835cab-f5ff-497c-a4c3-865e2a30f48f\") " pod="openstack/keystone-db-create-j5g9f" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.207655 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k4dv\" (UniqueName: \"kubernetes.io/projected/0c835cab-f5ff-497c-a4c3-865e2a30f48f-kube-api-access-8k4dv\") pod \"keystone-db-create-j5g9f\" (UID: \"0c835cab-f5ff-497c-a4c3-865e2a30f48f\") " pod="openstack/keystone-db-create-j5g9f" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.228281 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-f22wg"] Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.229218 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-f22wg" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.236894 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-f22wg"] Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.291409 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e-operator-scripts\") pod \"placement-db-create-f22wg\" (UID: \"dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e\") " pod="openstack/placement-db-create-f22wg" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.291471 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kzpg\" (UniqueName: \"kubernetes.io/projected/e4f446f4-2834-4ce5-8010-9e4308087ea1-kube-api-access-8kzpg\") pod \"keystone-9b47-account-create-update-8bqfg\" (UID: \"e4f446f4-2834-4ce5-8010-9e4308087ea1\") " pod="openstack/keystone-9b47-account-create-update-8bqfg" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.291499 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4f446f4-2834-4ce5-8010-9e4308087ea1-operator-scripts\") pod \"keystone-9b47-account-create-update-8bqfg\" (UID: \"e4f446f4-2834-4ce5-8010-9e4308087ea1\") " pod="openstack/keystone-9b47-account-create-update-8bqfg" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.291662 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26tdp\" (UniqueName: \"kubernetes.io/projected/dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e-kube-api-access-26tdp\") pod \"placement-db-create-f22wg\" (UID: \"dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e\") " pod="openstack/placement-db-create-f22wg" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.295597 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-64zxx" event={"ID":"110fcb3a-31c0-47d9-9e42-7e534e5ede02","Type":"ContainerDied","Data":"606c87698face0dead555133dafc8298d58f257c2d4b57407b02386aa764c43d"} Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.295628 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="606c87698face0dead555133dafc8298d58f257c2d4b57407b02386aa764c43d" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.295666 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-64zxx" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.338680 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-1631-account-create-update-mpg4t"] Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.339849 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1631-account-create-update-mpg4t" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.345378 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.347884 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-j5g9f" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.350851 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-1631-account-create-update-mpg4t"] Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.392678 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rbhl\" (UniqueName: \"kubernetes.io/projected/9dc075b4-dac3-46c3-9bad-9ffd34052643-kube-api-access-4rbhl\") pod \"placement-1631-account-create-update-mpg4t\" (UID: \"9dc075b4-dac3-46c3-9bad-9ffd34052643\") " pod="openstack/placement-1631-account-create-update-mpg4t" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.392749 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26tdp\" (UniqueName: \"kubernetes.io/projected/dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e-kube-api-access-26tdp\") pod \"placement-db-create-f22wg\" (UID: \"dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e\") " pod="openstack/placement-db-create-f22wg" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.392819 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9dc075b4-dac3-46c3-9bad-9ffd34052643-operator-scripts\") pod \"placement-1631-account-create-update-mpg4t\" (UID: \"9dc075b4-dac3-46c3-9bad-9ffd34052643\") " pod="openstack/placement-1631-account-create-update-mpg4t" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.392847 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e-operator-scripts\") pod \"placement-db-create-f22wg\" (UID: \"dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e\") " pod="openstack/placement-db-create-f22wg" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.392869 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kzpg\" (UniqueName: \"kubernetes.io/projected/e4f446f4-2834-4ce5-8010-9e4308087ea1-kube-api-access-8kzpg\") pod \"keystone-9b47-account-create-update-8bqfg\" (UID: \"e4f446f4-2834-4ce5-8010-9e4308087ea1\") " pod="openstack/keystone-9b47-account-create-update-8bqfg" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.392893 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4f446f4-2834-4ce5-8010-9e4308087ea1-operator-scripts\") pod \"keystone-9b47-account-create-update-8bqfg\" (UID: \"e4f446f4-2834-4ce5-8010-9e4308087ea1\") " pod="openstack/keystone-9b47-account-create-update-8bqfg" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.393577 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4f446f4-2834-4ce5-8010-9e4308087ea1-operator-scripts\") pod \"keystone-9b47-account-create-update-8bqfg\" (UID: \"e4f446f4-2834-4ce5-8010-9e4308087ea1\") " pod="openstack/keystone-9b47-account-create-update-8bqfg" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.393871 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e-operator-scripts\") pod \"placement-db-create-f22wg\" (UID: \"dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e\") " pod="openstack/placement-db-create-f22wg" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.416673 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26tdp\" (UniqueName: \"kubernetes.io/projected/dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e-kube-api-access-26tdp\") pod \"placement-db-create-f22wg\" (UID: \"dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e\") " pod="openstack/placement-db-create-f22wg" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.417112 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kzpg\" (UniqueName: \"kubernetes.io/projected/e4f446f4-2834-4ce5-8010-9e4308087ea1-kube-api-access-8kzpg\") pod \"keystone-9b47-account-create-update-8bqfg\" (UID: \"e4f446f4-2834-4ce5-8010-9e4308087ea1\") " pod="openstack/keystone-9b47-account-create-update-8bqfg" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.473149 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9b47-account-create-update-8bqfg" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.493993 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9dc075b4-dac3-46c3-9bad-9ffd34052643-operator-scripts\") pod \"placement-1631-account-create-update-mpg4t\" (UID: \"9dc075b4-dac3-46c3-9bad-9ffd34052643\") " pod="openstack/placement-1631-account-create-update-mpg4t" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.494446 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rbhl\" (UniqueName: \"kubernetes.io/projected/9dc075b4-dac3-46c3-9bad-9ffd34052643-kube-api-access-4rbhl\") pod \"placement-1631-account-create-update-mpg4t\" (UID: \"9dc075b4-dac3-46c3-9bad-9ffd34052643\") " pod="openstack/placement-1631-account-create-update-mpg4t" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.495456 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9dc075b4-dac3-46c3-9bad-9ffd34052643-operator-scripts\") pod \"placement-1631-account-create-update-mpg4t\" (UID: \"9dc075b4-dac3-46c3-9bad-9ffd34052643\") " pod="openstack/placement-1631-account-create-update-mpg4t" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.512613 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rbhl\" (UniqueName: \"kubernetes.io/projected/9dc075b4-dac3-46c3-9bad-9ffd34052643-kube-api-access-4rbhl\") pod \"placement-1631-account-create-update-mpg4t\" (UID: \"9dc075b4-dac3-46c3-9bad-9ffd34052643\") " pod="openstack/placement-1631-account-create-update-mpg4t" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.544163 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-f22wg" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.663626 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1631-account-create-update-mpg4t" Feb 27 16:42:30 crc kubenswrapper[4814]: I0227 16:42:30.799308 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:30 crc kubenswrapper[4814]: E0227 16:42:30.799499 4814 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 27 16:42:30 crc kubenswrapper[4814]: E0227 16:42:30.799546 4814 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 27 16:42:30 crc kubenswrapper[4814]: E0227 16:42:30.799625 4814 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift podName:9746d8ab-e2c8-419c-8327-6705260b31dd nodeName:}" failed. No retries permitted until 2026-02-27 16:42:38.799594134 +0000 UTC m=+1171.252218974 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift") pod "swift-storage-0" (UID: "9746d8ab-e2c8-419c-8327-6705260b31dd") : configmap "swift-ring-files" not found Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.104566 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.191087 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-kp4fh"] Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.191354 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" podUID="2368351c-7d12-4193-9d23-bac39e552df8" containerName="dnsmasq-dns" containerID="cri-o://c74323d36faf3ba14bc0582dd74bf9c49679b5c81d83a9dac533100e03ed07f4" gracePeriod=10 Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.286085 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-j5g9f"] Feb 27 16:42:32 crc kubenswrapper[4814]: W0227 16:42:32.292946 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a048b31_2392_4739_8350_93136b0820b3.slice/crio-646bf9ecc165daec47e03a7c1640f522a200e46798a97e2daa2e5cc2b67e1e64 WatchSource:0}: Error finding container 646bf9ecc165daec47e03a7c1640f522a200e46798a97e2daa2e5cc2b67e1e64: Status 404 returned error can't find the container with id 646bf9ecc165daec47e03a7c1640f522a200e46798a97e2daa2e5cc2b67e1e64 Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.296383 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-f22wg"] Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.317678 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8tfhv"] Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.319085 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-f22wg" event={"ID":"dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e","Type":"ContainerStarted","Data":"41246b80c6b376fdae4d7565de67cfbaa79cd3c94d97cf01a4a6b4296d4a374e"} Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.321421 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9481-account-create-update-rzblc" event={"ID":"653257dd-39a8-4062-b70f-33d78ef04baf","Type":"ContainerStarted","Data":"4288d55954c744b9f96668266e1e73400b0cbfb6985bbfe6f3ac3509614e7a66"} Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.323507 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-j5g9f" event={"ID":"0c835cab-f5ff-497c-a4c3-865e2a30f48f","Type":"ContainerStarted","Data":"9cc494e642298a9ce77720232547ee62e3a3a24625f22dedd8e1c3e817542c66"} Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.324865 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8tfhv" event={"ID":"7a048b31-2392-4739-8350-93136b0820b3","Type":"ContainerStarted","Data":"646bf9ecc165daec47e03a7c1640f522a200e46798a97e2daa2e5cc2b67e1e64"} Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.331185 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-k6ccx" event={"ID":"3daf276e-d557-4192-831c-d2aa124fe9bc","Type":"ContainerStarted","Data":"e967696a5b98f1272638155f0595f96ff0b7db65e4647f33bb94f0c6b565ba51"} Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.354462 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-9481-account-create-update-rzblc"] Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.375540 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-64zxx"] Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.380451 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-k6ccx" podStartSLOduration=2.350869116 podStartE2EDuration="6.380429769s" podCreationTimestamp="2026-02-27 16:42:26 +0000 UTC" firstStartedPulling="2026-02-27 16:42:27.706024402 +0000 UTC m=+1160.158649232" lastFinishedPulling="2026-02-27 16:42:31.735585045 +0000 UTC m=+1164.188209885" observedRunningTime="2026-02-27 16:42:32.350467088 +0000 UTC m=+1164.803091928" watchObservedRunningTime="2026-02-27 16:42:32.380429769 +0000 UTC m=+1164.833054599" Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.407462 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-64zxx"] Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.530364 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="110fcb3a-31c0-47d9-9e42-7e534e5ede02" path="/var/lib/kubelet/pods/110fcb3a-31c0-47d9-9e42-7e534e5ede02/volumes" Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.530900 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-1631-account-create-update-mpg4t"] Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.559689 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9b47-account-create-update-8bqfg"] Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.832007 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.959001 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2368351c-7d12-4193-9d23-bac39e552df8-dns-svc\") pod \"2368351c-7d12-4193-9d23-bac39e552df8\" (UID: \"2368351c-7d12-4193-9d23-bac39e552df8\") " Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.959143 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t8sc\" (UniqueName: \"kubernetes.io/projected/2368351c-7d12-4193-9d23-bac39e552df8-kube-api-access-4t8sc\") pod \"2368351c-7d12-4193-9d23-bac39e552df8\" (UID: \"2368351c-7d12-4193-9d23-bac39e552df8\") " Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.959217 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2368351c-7d12-4193-9d23-bac39e552df8-config\") pod \"2368351c-7d12-4193-9d23-bac39e552df8\" (UID: \"2368351c-7d12-4193-9d23-bac39e552df8\") " Feb 27 16:42:32 crc kubenswrapper[4814]: I0227 16:42:32.982834 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2368351c-7d12-4193-9d23-bac39e552df8-kube-api-access-4t8sc" (OuterVolumeSpecName: "kube-api-access-4t8sc") pod "2368351c-7d12-4193-9d23-bac39e552df8" (UID: "2368351c-7d12-4193-9d23-bac39e552df8"). InnerVolumeSpecName "kube-api-access-4t8sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.017044 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2368351c-7d12-4193-9d23-bac39e552df8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2368351c-7d12-4193-9d23-bac39e552df8" (UID: "2368351c-7d12-4193-9d23-bac39e552df8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.026496 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2368351c-7d12-4193-9d23-bac39e552df8-config" (OuterVolumeSpecName: "config") pod "2368351c-7d12-4193-9d23-bac39e552df8" (UID: "2368351c-7d12-4193-9d23-bac39e552df8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.060677 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2368351c-7d12-4193-9d23-bac39e552df8-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.060718 4814 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2368351c-7d12-4193-9d23-bac39e552df8-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.060731 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t8sc\" (UniqueName: \"kubernetes.io/projected/2368351c-7d12-4193-9d23-bac39e552df8-kube-api-access-4t8sc\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.342805 4814 generic.go:334] "Generic (PLEG): container finished" podID="7a048b31-2392-4739-8350-93136b0820b3" containerID="87966d0a1f0b1cd11e6e08bb7c206f16b3e7acd459379f5a938175794bbb9f11" exitCode=0 Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.342881 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8tfhv" event={"ID":"7a048b31-2392-4739-8350-93136b0820b3","Type":"ContainerDied","Data":"87966d0a1f0b1cd11e6e08bb7c206f16b3e7acd459379f5a938175794bbb9f11"} Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.346979 4814 generic.go:334] "Generic (PLEG): container finished" podID="e4f446f4-2834-4ce5-8010-9e4308087ea1" containerID="969868d9d157b53533159f7030e94a33751fb3bba169453cdd525ac1bdf44476" exitCode=0 Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.347040 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9b47-account-create-update-8bqfg" event={"ID":"e4f446f4-2834-4ce5-8010-9e4308087ea1","Type":"ContainerDied","Data":"969868d9d157b53533159f7030e94a33751fb3bba169453cdd525ac1bdf44476"} Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.347063 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9b47-account-create-update-8bqfg" event={"ID":"e4f446f4-2834-4ce5-8010-9e4308087ea1","Type":"ContainerStarted","Data":"9a28b007954e44ed878fbca824109dd3062f81f5aca7b40c5700e45428c84ae5"} Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.348951 4814 generic.go:334] "Generic (PLEG): container finished" podID="dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e" containerID="4a83fe3e3495cfc2556109aa00214204acd10f9e608f2528294960ea853e9c93" exitCode=0 Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.348998 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-f22wg" event={"ID":"dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e","Type":"ContainerDied","Data":"4a83fe3e3495cfc2556109aa00214204acd10f9e608f2528294960ea853e9c93"} Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.350762 4814 generic.go:334] "Generic (PLEG): container finished" podID="653257dd-39a8-4062-b70f-33d78ef04baf" containerID="51a1628bee460ac25f28940c3a4aa3bf5a67d97208f32aadf93d4d78850aa29b" exitCode=0 Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.350808 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9481-account-create-update-rzblc" event={"ID":"653257dd-39a8-4062-b70f-33d78ef04baf","Type":"ContainerDied","Data":"51a1628bee460ac25f28940c3a4aa3bf5a67d97208f32aadf93d4d78850aa29b"} Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.352946 4814 generic.go:334] "Generic (PLEG): container finished" podID="9dc075b4-dac3-46c3-9bad-9ffd34052643" containerID="abbee1a7af813844aae45f33cfda21df439528d9ef07c4277f524d0c02fca5cd" exitCode=0 Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.353009 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1631-account-create-update-mpg4t" event={"ID":"9dc075b4-dac3-46c3-9bad-9ffd34052643","Type":"ContainerDied","Data":"abbee1a7af813844aae45f33cfda21df439528d9ef07c4277f524d0c02fca5cd"} Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.353029 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1631-account-create-update-mpg4t" event={"ID":"9dc075b4-dac3-46c3-9bad-9ffd34052643","Type":"ContainerStarted","Data":"fe2320248127cbf10b1b233cfb60451487e535221f476d3dd6fdcb7a1339a4d9"} Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.358638 4814 generic.go:334] "Generic (PLEG): container finished" podID="2368351c-7d12-4193-9d23-bac39e552df8" containerID="c74323d36faf3ba14bc0582dd74bf9c49679b5c81d83a9dac533100e03ed07f4" exitCode=0 Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.358810 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.358828 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" event={"ID":"2368351c-7d12-4193-9d23-bac39e552df8","Type":"ContainerDied","Data":"c74323d36faf3ba14bc0582dd74bf9c49679b5c81d83a9dac533100e03ed07f4"} Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.358875 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-kp4fh" event={"ID":"2368351c-7d12-4193-9d23-bac39e552df8","Type":"ContainerDied","Data":"7e2a1f93711ab34b5078550750fecb272933e68acd51289390a4d45e465b5826"} Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.358902 4814 scope.go:117] "RemoveContainer" containerID="c74323d36faf3ba14bc0582dd74bf9c49679b5c81d83a9dac533100e03ed07f4" Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.367738 4814 generic.go:334] "Generic (PLEG): container finished" podID="0c835cab-f5ff-497c-a4c3-865e2a30f48f" containerID="c48c2203662b83108b76b2b3cdb1e68ae49384a83a7d8c64850261a730f4ca1d" exitCode=0 Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.367846 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-j5g9f" event={"ID":"0c835cab-f5ff-497c-a4c3-865e2a30f48f","Type":"ContainerDied","Data":"c48c2203662b83108b76b2b3cdb1e68ae49384a83a7d8c64850261a730f4ca1d"} Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.408290 4814 scope.go:117] "RemoveContainer" containerID="d27f66b53b70af6a019c1e53651b3c4e5acfe515cb3292df205015798e737991" Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.434323 4814 scope.go:117] "RemoveContainer" containerID="c74323d36faf3ba14bc0582dd74bf9c49679b5c81d83a9dac533100e03ed07f4" Feb 27 16:42:33 crc kubenswrapper[4814]: E0227 16:42:33.443796 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c74323d36faf3ba14bc0582dd74bf9c49679b5c81d83a9dac533100e03ed07f4\": container with ID starting with c74323d36faf3ba14bc0582dd74bf9c49679b5c81d83a9dac533100e03ed07f4 not found: ID does not exist" containerID="c74323d36faf3ba14bc0582dd74bf9c49679b5c81d83a9dac533100e03ed07f4" Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.443856 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c74323d36faf3ba14bc0582dd74bf9c49679b5c81d83a9dac533100e03ed07f4"} err="failed to get container status \"c74323d36faf3ba14bc0582dd74bf9c49679b5c81d83a9dac533100e03ed07f4\": rpc error: code = NotFound desc = could not find container \"c74323d36faf3ba14bc0582dd74bf9c49679b5c81d83a9dac533100e03ed07f4\": container with ID starting with c74323d36faf3ba14bc0582dd74bf9c49679b5c81d83a9dac533100e03ed07f4 not found: ID does not exist" Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.443927 4814 scope.go:117] "RemoveContainer" containerID="d27f66b53b70af6a019c1e53651b3c4e5acfe515cb3292df205015798e737991" Feb 27 16:42:33 crc kubenswrapper[4814]: E0227 16:42:33.444533 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d27f66b53b70af6a019c1e53651b3c4e5acfe515cb3292df205015798e737991\": container with ID starting with d27f66b53b70af6a019c1e53651b3c4e5acfe515cb3292df205015798e737991 not found: ID does not exist" containerID="d27f66b53b70af6a019c1e53651b3c4e5acfe515cb3292df205015798e737991" Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.444573 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d27f66b53b70af6a019c1e53651b3c4e5acfe515cb3292df205015798e737991"} err="failed to get container status \"d27f66b53b70af6a019c1e53651b3c4e5acfe515cb3292df205015798e737991\": rpc error: code = NotFound desc = could not find container \"d27f66b53b70af6a019c1e53651b3c4e5acfe515cb3292df205015798e737991\": container with ID starting with d27f66b53b70af6a019c1e53651b3c4e5acfe515cb3292df205015798e737991 not found: ID does not exist" Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.458839 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-kp4fh"] Feb 27 16:42:33 crc kubenswrapper[4814]: I0227 16:42:33.467461 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-kp4fh"] Feb 27 16:42:34 crc kubenswrapper[4814]: I0227 16:42:34.507655 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2368351c-7d12-4193-9d23-bac39e552df8" path="/var/lib/kubelet/pods/2368351c-7d12-4193-9d23-bac39e552df8/volumes" Feb 27 16:42:34 crc kubenswrapper[4814]: I0227 16:42:34.814858 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9b47-account-create-update-8bqfg" Feb 27 16:42:34 crc kubenswrapper[4814]: I0227 16:42:34.892722 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4f446f4-2834-4ce5-8010-9e4308087ea1-operator-scripts\") pod \"e4f446f4-2834-4ce5-8010-9e4308087ea1\" (UID: \"e4f446f4-2834-4ce5-8010-9e4308087ea1\") " Feb 27 16:42:34 crc kubenswrapper[4814]: I0227 16:42:34.892936 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kzpg\" (UniqueName: \"kubernetes.io/projected/e4f446f4-2834-4ce5-8010-9e4308087ea1-kube-api-access-8kzpg\") pod \"e4f446f4-2834-4ce5-8010-9e4308087ea1\" (UID: \"e4f446f4-2834-4ce5-8010-9e4308087ea1\") " Feb 27 16:42:34 crc kubenswrapper[4814]: I0227 16:42:34.894514 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4f446f4-2834-4ce5-8010-9e4308087ea1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e4f446f4-2834-4ce5-8010-9e4308087ea1" (UID: "e4f446f4-2834-4ce5-8010-9e4308087ea1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:34 crc kubenswrapper[4814]: I0227 16:42:34.900519 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4f446f4-2834-4ce5-8010-9e4308087ea1-kube-api-access-8kzpg" (OuterVolumeSpecName: "kube-api-access-8kzpg") pod "e4f446f4-2834-4ce5-8010-9e4308087ea1" (UID: "e4f446f4-2834-4ce5-8010-9e4308087ea1"). InnerVolumeSpecName "kube-api-access-8kzpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:42:34 crc kubenswrapper[4814]: I0227 16:42:34.994574 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kzpg\" (UniqueName: \"kubernetes.io/projected/e4f446f4-2834-4ce5-8010-9e4308087ea1-kube-api-access-8kzpg\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:34 crc kubenswrapper[4814]: I0227 16:42:34.994602 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4f446f4-2834-4ce5-8010-9e4308087ea1-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.010247 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8tfhv" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.018798 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-f22wg" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.027399 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1631-account-create-update-mpg4t" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.042550 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9481-account-create-update-rzblc" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.062035 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-j5g9f" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.095931 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9dc075b4-dac3-46c3-9bad-9ffd34052643-operator-scripts\") pod \"9dc075b4-dac3-46c3-9bad-9ffd34052643\" (UID: \"9dc075b4-dac3-46c3-9bad-9ffd34052643\") " Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.096009 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rbhl\" (UniqueName: \"kubernetes.io/projected/9dc075b4-dac3-46c3-9bad-9ffd34052643-kube-api-access-4rbhl\") pod \"9dc075b4-dac3-46c3-9bad-9ffd34052643\" (UID: \"9dc075b4-dac3-46c3-9bad-9ffd34052643\") " Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.096048 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e-operator-scripts\") pod \"dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e\" (UID: \"dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e\") " Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.096068 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26tdp\" (UniqueName: \"kubernetes.io/projected/dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e-kube-api-access-26tdp\") pod \"dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e\" (UID: \"dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e\") " Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.096090 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c835cab-f5ff-497c-a4c3-865e2a30f48f-operator-scripts\") pod \"0c835cab-f5ff-497c-a4c3-865e2a30f48f\" (UID: \"0c835cab-f5ff-497c-a4c3-865e2a30f48f\") " Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.096106 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8k4dv\" (UniqueName: \"kubernetes.io/projected/0c835cab-f5ff-497c-a4c3-865e2a30f48f-kube-api-access-8k4dv\") pod \"0c835cab-f5ff-497c-a4c3-865e2a30f48f\" (UID: \"0c835cab-f5ff-497c-a4c3-865e2a30f48f\") " Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.096157 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpnv6\" (UniqueName: \"kubernetes.io/projected/653257dd-39a8-4062-b70f-33d78ef04baf-kube-api-access-xpnv6\") pod \"653257dd-39a8-4062-b70f-33d78ef04baf\" (UID: \"653257dd-39a8-4062-b70f-33d78ef04baf\") " Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.096185 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a048b31-2392-4739-8350-93136b0820b3-operator-scripts\") pod \"7a048b31-2392-4739-8350-93136b0820b3\" (UID: \"7a048b31-2392-4739-8350-93136b0820b3\") " Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.096206 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skrcs\" (UniqueName: \"kubernetes.io/projected/7a048b31-2392-4739-8350-93136b0820b3-kube-api-access-skrcs\") pod \"7a048b31-2392-4739-8350-93136b0820b3\" (UID: \"7a048b31-2392-4739-8350-93136b0820b3\") " Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.096309 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/653257dd-39a8-4062-b70f-33d78ef04baf-operator-scripts\") pod \"653257dd-39a8-4062-b70f-33d78ef04baf\" (UID: \"653257dd-39a8-4062-b70f-33d78ef04baf\") " Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.097025 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/653257dd-39a8-4062-b70f-33d78ef04baf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "653257dd-39a8-4062-b70f-33d78ef04baf" (UID: "653257dd-39a8-4062-b70f-33d78ef04baf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.097390 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dc075b4-dac3-46c3-9bad-9ffd34052643-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9dc075b4-dac3-46c3-9bad-9ffd34052643" (UID: "9dc075b4-dac3-46c3-9bad-9ffd34052643"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.099282 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a048b31-2392-4739-8350-93136b0820b3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7a048b31-2392-4739-8350-93136b0820b3" (UID: "7a048b31-2392-4739-8350-93136b0820b3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.099417 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c835cab-f5ff-497c-a4c3-865e2a30f48f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0c835cab-f5ff-497c-a4c3-865e2a30f48f" (UID: "0c835cab-f5ff-497c-a4c3-865e2a30f48f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.100061 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e" (UID: "dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.101489 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c835cab-f5ff-497c-a4c3-865e2a30f48f-kube-api-access-8k4dv" (OuterVolumeSpecName: "kube-api-access-8k4dv") pod "0c835cab-f5ff-497c-a4c3-865e2a30f48f" (UID: "0c835cab-f5ff-497c-a4c3-865e2a30f48f"). InnerVolumeSpecName "kube-api-access-8k4dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.101521 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dc075b4-dac3-46c3-9bad-9ffd34052643-kube-api-access-4rbhl" (OuterVolumeSpecName: "kube-api-access-4rbhl") pod "9dc075b4-dac3-46c3-9bad-9ffd34052643" (UID: "9dc075b4-dac3-46c3-9bad-9ffd34052643"). InnerVolumeSpecName "kube-api-access-4rbhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.101934 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/653257dd-39a8-4062-b70f-33d78ef04baf-kube-api-access-xpnv6" (OuterVolumeSpecName: "kube-api-access-xpnv6") pod "653257dd-39a8-4062-b70f-33d78ef04baf" (UID: "653257dd-39a8-4062-b70f-33d78ef04baf"). InnerVolumeSpecName "kube-api-access-xpnv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.102800 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a048b31-2392-4739-8350-93136b0820b3-kube-api-access-skrcs" (OuterVolumeSpecName: "kube-api-access-skrcs") pod "7a048b31-2392-4739-8350-93136b0820b3" (UID: "7a048b31-2392-4739-8350-93136b0820b3"). InnerVolumeSpecName "kube-api-access-skrcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.103366 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e-kube-api-access-26tdp" (OuterVolumeSpecName: "kube-api-access-26tdp") pod "dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e" (UID: "dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e"). InnerVolumeSpecName "kube-api-access-26tdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.198660 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26tdp\" (UniqueName: \"kubernetes.io/projected/dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e-kube-api-access-26tdp\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.198705 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c835cab-f5ff-497c-a4c3-865e2a30f48f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.198715 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8k4dv\" (UniqueName: \"kubernetes.io/projected/0c835cab-f5ff-497c-a4c3-865e2a30f48f-kube-api-access-8k4dv\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.198724 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpnv6\" (UniqueName: \"kubernetes.io/projected/653257dd-39a8-4062-b70f-33d78ef04baf-kube-api-access-xpnv6\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.198733 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a048b31-2392-4739-8350-93136b0820b3-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.198741 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skrcs\" (UniqueName: \"kubernetes.io/projected/7a048b31-2392-4739-8350-93136b0820b3-kube-api-access-skrcs\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.198748 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/653257dd-39a8-4062-b70f-33d78ef04baf-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.198760 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9dc075b4-dac3-46c3-9bad-9ffd34052643-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.198768 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rbhl\" (UniqueName: \"kubernetes.io/projected/9dc075b4-dac3-46c3-9bad-9ffd34052643-kube-api-access-4rbhl\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.198777 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.387499 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-j5g9f" event={"ID":"0c835cab-f5ff-497c-a4c3-865e2a30f48f","Type":"ContainerDied","Data":"9cc494e642298a9ce77720232547ee62e3a3a24625f22dedd8e1c3e817542c66"} Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.387491 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-j5g9f" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.387565 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cc494e642298a9ce77720232547ee62e3a3a24625f22dedd8e1c3e817542c66" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.389014 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8tfhv" event={"ID":"7a048b31-2392-4739-8350-93136b0820b3","Type":"ContainerDied","Data":"646bf9ecc165daec47e03a7c1640f522a200e46798a97e2daa2e5cc2b67e1e64"} Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.389046 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8tfhv" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.389053 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="646bf9ecc165daec47e03a7c1640f522a200e46798a97e2daa2e5cc2b67e1e64" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.390359 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9b47-account-create-update-8bqfg" event={"ID":"e4f446f4-2834-4ce5-8010-9e4308087ea1","Type":"ContainerDied","Data":"9a28b007954e44ed878fbca824109dd3062f81f5aca7b40c5700e45428c84ae5"} Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.390382 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a28b007954e44ed878fbca824109dd3062f81f5aca7b40c5700e45428c84ae5" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.390445 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9b47-account-create-update-8bqfg" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.397419 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-f22wg" event={"ID":"dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e","Type":"ContainerDied","Data":"41246b80c6b376fdae4d7565de67cfbaa79cd3c94d97cf01a4a6b4296d4a374e"} Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.397461 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41246b80c6b376fdae4d7565de67cfbaa79cd3c94d97cf01a4a6b4296d4a374e" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.397523 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-f22wg" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.402084 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9481-account-create-update-rzblc" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.402092 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9481-account-create-update-rzblc" event={"ID":"653257dd-39a8-4062-b70f-33d78ef04baf","Type":"ContainerDied","Data":"4288d55954c744b9f96668266e1e73400b0cbfb6985bbfe6f3ac3509614e7a66"} Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.402200 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4288d55954c744b9f96668266e1e73400b0cbfb6985bbfe6f3ac3509614e7a66" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.404044 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1631-account-create-update-mpg4t" event={"ID":"9dc075b4-dac3-46c3-9bad-9ffd34052643","Type":"ContainerDied","Data":"fe2320248127cbf10b1b233cfb60451487e535221f476d3dd6fdcb7a1339a4d9"} Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.404090 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe2320248127cbf10b1b233cfb60451487e535221f476d3dd6fdcb7a1339a4d9" Feb 27 16:42:35 crc kubenswrapper[4814]: I0227 16:42:35.404110 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1631-account-create-update-mpg4t" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.375761 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-ln7b6"] Feb 27 16:42:37 crc kubenswrapper[4814]: E0227 16:42:37.378498 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dc075b4-dac3-46c3-9bad-9ffd34052643" containerName="mariadb-account-create-update" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.378604 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dc075b4-dac3-46c3-9bad-9ffd34052643" containerName="mariadb-account-create-update" Feb 27 16:42:37 crc kubenswrapper[4814]: E0227 16:42:37.378693 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c835cab-f5ff-497c-a4c3-865e2a30f48f" containerName="mariadb-database-create" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.378768 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c835cab-f5ff-497c-a4c3-865e2a30f48f" containerName="mariadb-database-create" Feb 27 16:42:37 crc kubenswrapper[4814]: E0227 16:42:37.378866 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="653257dd-39a8-4062-b70f-33d78ef04baf" containerName="mariadb-account-create-update" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.378957 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="653257dd-39a8-4062-b70f-33d78ef04baf" containerName="mariadb-account-create-update" Feb 27 16:42:37 crc kubenswrapper[4814]: E0227 16:42:37.379065 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2368351c-7d12-4193-9d23-bac39e552df8" containerName="init" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.379833 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2368351c-7d12-4193-9d23-bac39e552df8" containerName="init" Feb 27 16:42:37 crc kubenswrapper[4814]: E0227 16:42:37.379961 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f446f4-2834-4ce5-8010-9e4308087ea1" containerName="mariadb-account-create-update" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.380123 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f446f4-2834-4ce5-8010-9e4308087ea1" containerName="mariadb-account-create-update" Feb 27 16:42:37 crc kubenswrapper[4814]: E0227 16:42:37.380282 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a048b31-2392-4739-8350-93136b0820b3" containerName="mariadb-database-create" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.380420 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a048b31-2392-4739-8350-93136b0820b3" containerName="mariadb-database-create" Feb 27 16:42:37 crc kubenswrapper[4814]: E0227 16:42:37.380569 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2368351c-7d12-4193-9d23-bac39e552df8" containerName="dnsmasq-dns" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.380701 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2368351c-7d12-4193-9d23-bac39e552df8" containerName="dnsmasq-dns" Feb 27 16:42:37 crc kubenswrapper[4814]: E0227 16:42:37.380806 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e" containerName="mariadb-database-create" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.380884 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e" containerName="mariadb-database-create" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.381191 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f446f4-2834-4ce5-8010-9e4308087ea1" containerName="mariadb-account-create-update" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.381308 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dc075b4-dac3-46c3-9bad-9ffd34052643" containerName="mariadb-account-create-update" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.381419 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e" containerName="mariadb-database-create" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.381837 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a048b31-2392-4739-8350-93136b0820b3" containerName="mariadb-database-create" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.382028 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c835cab-f5ff-497c-a4c3-865e2a30f48f" containerName="mariadb-database-create" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.382115 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="653257dd-39a8-4062-b70f-33d78ef04baf" containerName="mariadb-account-create-update" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.382215 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="2368351c-7d12-4193-9d23-bac39e552df8" containerName="dnsmasq-dns" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.383286 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-ln7b6" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.386208 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.401688 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-ln7b6"] Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.442492 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64ft4\" (UniqueName: \"kubernetes.io/projected/eaef6a7d-1318-40a2-a11a-2e9e1241f828-kube-api-access-64ft4\") pod \"root-account-create-update-ln7b6\" (UID: \"eaef6a7d-1318-40a2-a11a-2e9e1241f828\") " pod="openstack/root-account-create-update-ln7b6" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.442623 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eaef6a7d-1318-40a2-a11a-2e9e1241f828-operator-scripts\") pod \"root-account-create-update-ln7b6\" (UID: \"eaef6a7d-1318-40a2-a11a-2e9e1241f828\") " pod="openstack/root-account-create-update-ln7b6" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.544620 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eaef6a7d-1318-40a2-a11a-2e9e1241f828-operator-scripts\") pod \"root-account-create-update-ln7b6\" (UID: \"eaef6a7d-1318-40a2-a11a-2e9e1241f828\") " pod="openstack/root-account-create-update-ln7b6" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.544802 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64ft4\" (UniqueName: \"kubernetes.io/projected/eaef6a7d-1318-40a2-a11a-2e9e1241f828-kube-api-access-64ft4\") pod \"root-account-create-update-ln7b6\" (UID: \"eaef6a7d-1318-40a2-a11a-2e9e1241f828\") " pod="openstack/root-account-create-update-ln7b6" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.546728 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eaef6a7d-1318-40a2-a11a-2e9e1241f828-operator-scripts\") pod \"root-account-create-update-ln7b6\" (UID: \"eaef6a7d-1318-40a2-a11a-2e9e1241f828\") " pod="openstack/root-account-create-update-ln7b6" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.570538 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64ft4\" (UniqueName: \"kubernetes.io/projected/eaef6a7d-1318-40a2-a11a-2e9e1241f828-kube-api-access-64ft4\") pod \"root-account-create-update-ln7b6\" (UID: \"eaef6a7d-1318-40a2-a11a-2e9e1241f828\") " pod="openstack/root-account-create-update-ln7b6" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.714299 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-ln7b6" Feb 27 16:42:37 crc kubenswrapper[4814]: I0227 16:42:37.931044 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 27 16:42:38 crc kubenswrapper[4814]: I0227 16:42:38.220629 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-ln7b6"] Feb 27 16:42:38 crc kubenswrapper[4814]: I0227 16:42:38.429698 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-ln7b6" event={"ID":"eaef6a7d-1318-40a2-a11a-2e9e1241f828","Type":"ContainerStarted","Data":"f8f054552a5c11789a93dc8da2f52e6207e11331d4008eb36ea8a980b12ce547"} Feb 27 16:42:38 crc kubenswrapper[4814]: I0227 16:42:38.429788 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-ln7b6" event={"ID":"eaef6a7d-1318-40a2-a11a-2e9e1241f828","Type":"ContainerStarted","Data":"625cf89b59c8d36c1b1bcf9c3c3735dc2b94af054267ef8822f4a609692d8780"} Feb 27 16:42:38 crc kubenswrapper[4814]: I0227 16:42:38.443528 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-ln7b6" podStartSLOduration=1.443486716 podStartE2EDuration="1.443486716s" podCreationTimestamp="2026-02-27 16:42:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:42:38.442193505 +0000 UTC m=+1170.894818335" watchObservedRunningTime="2026-02-27 16:42:38.443486716 +0000 UTC m=+1170.896111556" Feb 27 16:42:38 crc kubenswrapper[4814]: I0227 16:42:38.884415 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:38 crc kubenswrapper[4814]: I0227 16:42:38.897519 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9746d8ab-e2c8-419c-8327-6705260b31dd-etc-swift\") pod \"swift-storage-0\" (UID: \"9746d8ab-e2c8-419c-8327-6705260b31dd\") " pod="openstack/swift-storage-0" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.161801 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.432434 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-k85vk"] Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.434162 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-k85vk" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.437825 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-dgrwq" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.438042 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.441336 4814 generic.go:334] "Generic (PLEG): container finished" podID="3daf276e-d557-4192-831c-d2aa124fe9bc" containerID="e967696a5b98f1272638155f0595f96ff0b7db65e4647f33bb94f0c6b565ba51" exitCode=0 Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.441425 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-k6ccx" event={"ID":"3daf276e-d557-4192-831c-d2aa124fe9bc","Type":"ContainerDied","Data":"e967696a5b98f1272638155f0595f96ff0b7db65e4647f33bb94f0c6b565ba51"} Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.443439 4814 generic.go:334] "Generic (PLEG): container finished" podID="eaef6a7d-1318-40a2-a11a-2e9e1241f828" containerID="f8f054552a5c11789a93dc8da2f52e6207e11331d4008eb36ea8a980b12ce547" exitCode=0 Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.443467 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-ln7b6" event={"ID":"eaef6a7d-1318-40a2-a11a-2e9e1241f828","Type":"ContainerDied","Data":"f8f054552a5c11789a93dc8da2f52e6207e11331d4008eb36ea8a980b12ce547"} Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.444661 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-k85vk"] Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.499721 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-combined-ca-bundle\") pod \"glance-db-sync-k85vk\" (UID: \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\") " pod="openstack/glance-db-sync-k85vk" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.499823 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-config-data\") pod \"glance-db-sync-k85vk\" (UID: \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\") " pod="openstack/glance-db-sync-k85vk" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.499913 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cnfr\" (UniqueName: \"kubernetes.io/projected/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-kube-api-access-7cnfr\") pod \"glance-db-sync-k85vk\" (UID: \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\") " pod="openstack/glance-db-sync-k85vk" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.499977 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-db-sync-config-data\") pod \"glance-db-sync-k85vk\" (UID: \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\") " pod="openstack/glance-db-sync-k85vk" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.540951 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-9tkjl" podUID="56c74238-2dde-4709-922a-9551da5fe8ae" containerName="ovn-controller" probeResult="failure" output=< Feb 27 16:42:39 crc kubenswrapper[4814]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 27 16:42:39 crc kubenswrapper[4814]: > Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.601953 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-db-sync-config-data\") pod \"glance-db-sync-k85vk\" (UID: \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\") " pod="openstack/glance-db-sync-k85vk" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.602054 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-combined-ca-bundle\") pod \"glance-db-sync-k85vk\" (UID: \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\") " pod="openstack/glance-db-sync-k85vk" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.602115 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-config-data\") pod \"glance-db-sync-k85vk\" (UID: \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\") " pod="openstack/glance-db-sync-k85vk" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.602176 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cnfr\" (UniqueName: \"kubernetes.io/projected/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-kube-api-access-7cnfr\") pod \"glance-db-sync-k85vk\" (UID: \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\") " pod="openstack/glance-db-sync-k85vk" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.608401 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-combined-ca-bundle\") pod \"glance-db-sync-k85vk\" (UID: \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\") " pod="openstack/glance-db-sync-k85vk" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.610878 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-config-data\") pod \"glance-db-sync-k85vk\" (UID: \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\") " pod="openstack/glance-db-sync-k85vk" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.622246 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-db-sync-config-data\") pod \"glance-db-sync-k85vk\" (UID: \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\") " pod="openstack/glance-db-sync-k85vk" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.628093 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cnfr\" (UniqueName: \"kubernetes.io/projected/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-kube-api-access-7cnfr\") pod \"glance-db-sync-k85vk\" (UID: \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\") " pod="openstack/glance-db-sync-k85vk" Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.753038 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 27 16:42:39 crc kubenswrapper[4814]: W0227 16:42:39.756085 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9746d8ab_e2c8_419c_8327_6705260b31dd.slice/crio-d594a7585a7b6df6041e9c4286837b88d30baba4b4f74016cee6dae9b0613168 WatchSource:0}: Error finding container d594a7585a7b6df6041e9c4286837b88d30baba4b4f74016cee6dae9b0613168: Status 404 returned error can't find the container with id d594a7585a7b6df6041e9c4286837b88d30baba4b4f74016cee6dae9b0613168 Feb 27 16:42:39 crc kubenswrapper[4814]: I0227 16:42:39.758444 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-k85vk" Feb 27 16:42:40 crc kubenswrapper[4814]: I0227 16:42:40.342020 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-k85vk"] Feb 27 16:42:40 crc kubenswrapper[4814]: I0227 16:42:40.452758 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-k85vk" event={"ID":"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9","Type":"ContainerStarted","Data":"5a843f21106d4575e4dcea76ea8f25fea010552528e7252f3649804690d91137"} Feb 27 16:42:40 crc kubenswrapper[4814]: I0227 16:42:40.454219 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9746d8ab-e2c8-419c-8327-6705260b31dd","Type":"ContainerStarted","Data":"d594a7585a7b6df6041e9c4286837b88d30baba4b4f74016cee6dae9b0613168"} Feb 27 16:42:40 crc kubenswrapper[4814]: I0227 16:42:40.919041 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-ln7b6" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.006589 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.031434 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64ft4\" (UniqueName: \"kubernetes.io/projected/eaef6a7d-1318-40a2-a11a-2e9e1241f828-kube-api-access-64ft4\") pod \"eaef6a7d-1318-40a2-a11a-2e9e1241f828\" (UID: \"eaef6a7d-1318-40a2-a11a-2e9e1241f828\") " Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.031600 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eaef6a7d-1318-40a2-a11a-2e9e1241f828-operator-scripts\") pod \"eaef6a7d-1318-40a2-a11a-2e9e1241f828\" (UID: \"eaef6a7d-1318-40a2-a11a-2e9e1241f828\") " Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.032807 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaef6a7d-1318-40a2-a11a-2e9e1241f828-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eaef6a7d-1318-40a2-a11a-2e9e1241f828" (UID: "eaef6a7d-1318-40a2-a11a-2e9e1241f828"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.040622 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaef6a7d-1318-40a2-a11a-2e9e1241f828-kube-api-access-64ft4" (OuterVolumeSpecName: "kube-api-access-64ft4") pod "eaef6a7d-1318-40a2-a11a-2e9e1241f828" (UID: "eaef6a7d-1318-40a2-a11a-2e9e1241f828"). InnerVolumeSpecName "kube-api-access-64ft4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.133489 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3daf276e-d557-4192-831c-d2aa124fe9bc-ring-data-devices\") pod \"3daf276e-d557-4192-831c-d2aa124fe9bc\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.133589 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-swiftconf\") pod \"3daf276e-d557-4192-831c-d2aa124fe9bc\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.133682 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-combined-ca-bundle\") pod \"3daf276e-d557-4192-831c-d2aa124fe9bc\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.133729 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-dispersionconf\") pod \"3daf276e-d557-4192-831c-d2aa124fe9bc\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.133757 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsnl2\" (UniqueName: \"kubernetes.io/projected/3daf276e-d557-4192-831c-d2aa124fe9bc-kube-api-access-tsnl2\") pod \"3daf276e-d557-4192-831c-d2aa124fe9bc\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.133800 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3daf276e-d557-4192-831c-d2aa124fe9bc-etc-swift\") pod \"3daf276e-d557-4192-831c-d2aa124fe9bc\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.133887 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3daf276e-d557-4192-831c-d2aa124fe9bc-scripts\") pod \"3daf276e-d557-4192-831c-d2aa124fe9bc\" (UID: \"3daf276e-d557-4192-831c-d2aa124fe9bc\") " Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.134312 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64ft4\" (UniqueName: \"kubernetes.io/projected/eaef6a7d-1318-40a2-a11a-2e9e1241f828-kube-api-access-64ft4\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.134340 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eaef6a7d-1318-40a2-a11a-2e9e1241f828-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.135049 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3daf276e-d557-4192-831c-d2aa124fe9bc-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3daf276e-d557-4192-831c-d2aa124fe9bc" (UID: "3daf276e-d557-4192-831c-d2aa124fe9bc"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.135241 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3daf276e-d557-4192-831c-d2aa124fe9bc-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3daf276e-d557-4192-831c-d2aa124fe9bc" (UID: "3daf276e-d557-4192-831c-d2aa124fe9bc"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.140642 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3daf276e-d557-4192-831c-d2aa124fe9bc-kube-api-access-tsnl2" (OuterVolumeSpecName: "kube-api-access-tsnl2") pod "3daf276e-d557-4192-831c-d2aa124fe9bc" (UID: "3daf276e-d557-4192-831c-d2aa124fe9bc"). InnerVolumeSpecName "kube-api-access-tsnl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.152476 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3daf276e-d557-4192-831c-d2aa124fe9bc" (UID: "3daf276e-d557-4192-831c-d2aa124fe9bc"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.157392 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3daf276e-d557-4192-831c-d2aa124fe9bc" (UID: "3daf276e-d557-4192-831c-d2aa124fe9bc"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.160616 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3daf276e-d557-4192-831c-d2aa124fe9bc" (UID: "3daf276e-d557-4192-831c-d2aa124fe9bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.170285 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3daf276e-d557-4192-831c-d2aa124fe9bc-scripts" (OuterVolumeSpecName: "scripts") pod "3daf276e-d557-4192-831c-d2aa124fe9bc" (UID: "3daf276e-d557-4192-831c-d2aa124fe9bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.237654 4814 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.237858 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.237872 4814 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3daf276e-d557-4192-831c-d2aa124fe9bc-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.237911 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsnl2\" (UniqueName: \"kubernetes.io/projected/3daf276e-d557-4192-831c-d2aa124fe9bc-kube-api-access-tsnl2\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.237926 4814 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3daf276e-d557-4192-831c-d2aa124fe9bc-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.237939 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3daf276e-d557-4192-831c-d2aa124fe9bc-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.237950 4814 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3daf276e-d557-4192-831c-d2aa124fe9bc-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.473067 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-ln7b6" event={"ID":"eaef6a7d-1318-40a2-a11a-2e9e1241f828","Type":"ContainerDied","Data":"625cf89b59c8d36c1b1bcf9c3c3735dc2b94af054267ef8822f4a609692d8780"} Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.473094 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-ln7b6" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.473129 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="625cf89b59c8d36c1b1bcf9c3c3735dc2b94af054267ef8822f4a609692d8780" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.479939 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9746d8ab-e2c8-419c-8327-6705260b31dd","Type":"ContainerStarted","Data":"f2a461960e629237cf879c7b2cf603114159e57a0eda2bac7becea829dc031cf"} Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.479991 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9746d8ab-e2c8-419c-8327-6705260b31dd","Type":"ContainerStarted","Data":"2bb79e594d72e9a55cd22d90b1f09f68cadabfd723682443f3b15dd41fb35dff"} Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.481681 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-k6ccx" event={"ID":"3daf276e-d557-4192-831c-d2aa124fe9bc","Type":"ContainerDied","Data":"8dc718481d252c2a39c32c2c35abea68932b1953547dce0055be1518a92d15ba"} Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.481705 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dc718481d252c2a39c32c2c35abea68932b1953547dce0055be1518a92d15ba" Feb 27 16:42:41 crc kubenswrapper[4814]: I0227 16:42:41.481791 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-k6ccx" Feb 27 16:42:42 crc kubenswrapper[4814]: I0227 16:42:42.498445 4814 generic.go:334] "Generic (PLEG): container finished" podID="58660097-6b23-4016-98b4-6a10978c0887" containerID="3901d5b87aa0457909d2bb6c3ba5880bef5442d1430009b012abf06b49f82fba" exitCode=0 Feb 27 16:42:42 crc kubenswrapper[4814]: I0227 16:42:42.501966 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"58660097-6b23-4016-98b4-6a10978c0887","Type":"ContainerDied","Data":"3901d5b87aa0457909d2bb6c3ba5880bef5442d1430009b012abf06b49f82fba"} Feb 27 16:42:42 crc kubenswrapper[4814]: I0227 16:42:42.504505 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9746d8ab-e2c8-419c-8327-6705260b31dd","Type":"ContainerStarted","Data":"5f4db086be6162f1dda8c4904f1b1e6f59a170e567aaceca518a807530e961ea"} Feb 27 16:42:42 crc kubenswrapper[4814]: I0227 16:42:42.504532 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9746d8ab-e2c8-419c-8327-6705260b31dd","Type":"ContainerStarted","Data":"73553763328c04374b738a9293aa249e401443203b4bc5d2fae7a257ab76055c"} Feb 27 16:42:43 crc kubenswrapper[4814]: I0227 16:42:43.518981 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"58660097-6b23-4016-98b4-6a10978c0887","Type":"ContainerStarted","Data":"a535cf2aaf66daf9350afc5bbf95a9a881e393aac3022afed976df8d33d602f2"} Feb 27 16:42:43 crc kubenswrapper[4814]: I0227 16:42:43.519298 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 27 16:42:43 crc kubenswrapper[4814]: I0227 16:42:43.521850 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9746d8ab-e2c8-419c-8327-6705260b31dd","Type":"ContainerStarted","Data":"639b3b2ab1d3e47c0c16d7717b7606a4272105bc44089daf198feab276f7a58f"} Feb 27 16:42:43 crc kubenswrapper[4814]: I0227 16:42:43.521878 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9746d8ab-e2c8-419c-8327-6705260b31dd","Type":"ContainerStarted","Data":"9f9577543cc673e1399ed8d54fc8fc0ec51335cd1050a2f544254f2d39e884c3"} Feb 27 16:42:43 crc kubenswrapper[4814]: I0227 16:42:43.522954 4814 generic.go:334] "Generic (PLEG): container finished" podID="8bfc5fb5-c9ab-4108-915b-e6669b332e05" containerID="2c129c7f4218bd10e27a0ca4ecd6c565a206b1a457efa3f4a5f5c065833f539d" exitCode=0 Feb 27 16:42:43 crc kubenswrapper[4814]: I0227 16:42:43.522991 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8bfc5fb5-c9ab-4108-915b-e6669b332e05","Type":"ContainerDied","Data":"2c129c7f4218bd10e27a0ca4ecd6c565a206b1a457efa3f4a5f5c065833f539d"} Feb 27 16:42:43 crc kubenswrapper[4814]: I0227 16:42:43.592352 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=50.876311365 podStartE2EDuration="59.592331811s" podCreationTimestamp="2026-02-27 16:41:44 +0000 UTC" firstStartedPulling="2026-02-27 16:41:59.197571563 +0000 UTC m=+1131.650196383" lastFinishedPulling="2026-02-27 16:42:07.913591999 +0000 UTC m=+1140.366216829" observedRunningTime="2026-02-27 16:42:43.580354795 +0000 UTC m=+1176.032979625" watchObservedRunningTime="2026-02-27 16:42:43.592331811 +0000 UTC m=+1176.044956761" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.540131 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-9tkjl" podUID="56c74238-2dde-4709-922a-9551da5fe8ae" containerName="ovn-controller" probeResult="failure" output=< Feb 27 16:42:44 crc kubenswrapper[4814]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 27 16:42:44 crc kubenswrapper[4814]: > Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.549529 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8bfc5fb5-c9ab-4108-915b-e6669b332e05","Type":"ContainerStarted","Data":"0375689c55064914df1dda469516539c22548c73611fd0971ce7f5ec9e6fe3a0"} Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.549743 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.557117 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9746d8ab-e2c8-419c-8327-6705260b31dd","Type":"ContainerStarted","Data":"d092da36fc739b950d2fe1292051fdf25ee5c15954e3ecc58660248787b41d12"} Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.557166 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9746d8ab-e2c8-419c-8327-6705260b31dd","Type":"ContainerStarted","Data":"4f45b8591ce6b8345b795f380442494b1002d05c5671a822f7a436a58a3eeb3d"} Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.560740 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.564645 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-kxqxw" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.582060 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=52.076641722 podStartE2EDuration="1m0.582044298s" podCreationTimestamp="2026-02-27 16:41:44 +0000 UTC" firstStartedPulling="2026-02-27 16:41:59.538228825 +0000 UTC m=+1131.990853655" lastFinishedPulling="2026-02-27 16:42:08.043631401 +0000 UTC m=+1140.496256231" observedRunningTime="2026-02-27 16:42:44.578426695 +0000 UTC m=+1177.031051535" watchObservedRunningTime="2026-02-27 16:42:44.582044298 +0000 UTC m=+1177.034669138" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.802774 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-9tkjl-config-l7jb5"] Feb 27 16:42:44 crc kubenswrapper[4814]: E0227 16:42:44.803842 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaef6a7d-1318-40a2-a11a-2e9e1241f828" containerName="mariadb-account-create-update" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.803863 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaef6a7d-1318-40a2-a11a-2e9e1241f828" containerName="mariadb-account-create-update" Feb 27 16:42:44 crc kubenswrapper[4814]: E0227 16:42:44.803873 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3daf276e-d557-4192-831c-d2aa124fe9bc" containerName="swift-ring-rebalance" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.803880 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="3daf276e-d557-4192-831c-d2aa124fe9bc" containerName="swift-ring-rebalance" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.804600 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="3daf276e-d557-4192-831c-d2aa124fe9bc" containerName="swift-ring-rebalance" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.804629 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaef6a7d-1318-40a2-a11a-2e9e1241f828" containerName="mariadb-account-create-update" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.805384 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.811619 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.817870 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9tkjl-config-l7jb5"] Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.917713 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-run\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.917768 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-run-ovn\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.917795 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c50f3d11-a17a-4e29-be10-ed31d81163a2-scripts\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.917857 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c50f3d11-a17a-4e29-be10-ed31d81163a2-additional-scripts\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.917954 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vh7s7\" (UniqueName: \"kubernetes.io/projected/c50f3d11-a17a-4e29-be10-ed31d81163a2-kube-api-access-vh7s7\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:44 crc kubenswrapper[4814]: I0227 16:42:44.917984 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-log-ovn\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:45 crc kubenswrapper[4814]: I0227 16:42:45.019641 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vh7s7\" (UniqueName: \"kubernetes.io/projected/c50f3d11-a17a-4e29-be10-ed31d81163a2-kube-api-access-vh7s7\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:45 crc kubenswrapper[4814]: I0227 16:42:45.019723 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-log-ovn\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:45 crc kubenswrapper[4814]: I0227 16:42:45.019757 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-run\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:45 crc kubenswrapper[4814]: I0227 16:42:45.019804 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-run-ovn\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:45 crc kubenswrapper[4814]: I0227 16:42:45.019853 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c50f3d11-a17a-4e29-be10-ed31d81163a2-scripts\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:45 crc kubenswrapper[4814]: I0227 16:42:45.019879 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c50f3d11-a17a-4e29-be10-ed31d81163a2-additional-scripts\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:45 crc kubenswrapper[4814]: I0227 16:42:45.020103 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-log-ovn\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:45 crc kubenswrapper[4814]: I0227 16:42:45.020162 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-run\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:45 crc kubenswrapper[4814]: I0227 16:42:45.020182 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-run-ovn\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:45 crc kubenswrapper[4814]: I0227 16:42:45.021468 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c50f3d11-a17a-4e29-be10-ed31d81163a2-additional-scripts\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:45 crc kubenswrapper[4814]: I0227 16:42:45.022210 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c50f3d11-a17a-4e29-be10-ed31d81163a2-scripts\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:45 crc kubenswrapper[4814]: I0227 16:42:45.055104 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vh7s7\" (UniqueName: \"kubernetes.io/projected/c50f3d11-a17a-4e29-be10-ed31d81163a2-kube-api-access-vh7s7\") pod \"ovn-controller-9tkjl-config-l7jb5\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:45 crc kubenswrapper[4814]: I0227 16:42:45.132537 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:45 crc kubenswrapper[4814]: I0227 16:42:45.659630 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9tkjl-config-l7jb5"] Feb 27 16:42:46 crc kubenswrapper[4814]: I0227 16:42:46.574920 4814 generic.go:334] "Generic (PLEG): container finished" podID="c50f3d11-a17a-4e29-be10-ed31d81163a2" containerID="ba75ebdc4923da74d710b0b62d701e35a031dd4a66ec36db84223e7fdbd063ea" exitCode=0 Feb 27 16:42:46 crc kubenswrapper[4814]: I0227 16:42:46.575025 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9tkjl-config-l7jb5" event={"ID":"c50f3d11-a17a-4e29-be10-ed31d81163a2","Type":"ContainerDied","Data":"ba75ebdc4923da74d710b0b62d701e35a031dd4a66ec36db84223e7fdbd063ea"} Feb 27 16:42:46 crc kubenswrapper[4814]: I0227 16:42:46.575187 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9tkjl-config-l7jb5" event={"ID":"c50f3d11-a17a-4e29-be10-ed31d81163a2","Type":"ContainerStarted","Data":"1c039d0fce03df46f36646593980ae82e9385c0071576ae497bd0fbc84a54a58"} Feb 27 16:42:49 crc kubenswrapper[4814]: I0227 16:42:49.539394 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-9tkjl" Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.154644 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.287696 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-run-ovn\") pod \"c50f3d11-a17a-4e29-be10-ed31d81163a2\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.287761 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-log-ovn\") pod \"c50f3d11-a17a-4e29-be10-ed31d81163a2\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.287816 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-run\") pod \"c50f3d11-a17a-4e29-be10-ed31d81163a2\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.287873 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c50f3d11-a17a-4e29-be10-ed31d81163a2-scripts\") pod \"c50f3d11-a17a-4e29-be10-ed31d81163a2\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.287905 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c50f3d11-a17a-4e29-be10-ed31d81163a2-additional-scripts\") pod \"c50f3d11-a17a-4e29-be10-ed31d81163a2\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.287925 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vh7s7\" (UniqueName: \"kubernetes.io/projected/c50f3d11-a17a-4e29-be10-ed31d81163a2-kube-api-access-vh7s7\") pod \"c50f3d11-a17a-4e29-be10-ed31d81163a2\" (UID: \"c50f3d11-a17a-4e29-be10-ed31d81163a2\") " Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.288422 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "c50f3d11-a17a-4e29-be10-ed31d81163a2" (UID: "c50f3d11-a17a-4e29-be10-ed31d81163a2"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.288468 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "c50f3d11-a17a-4e29-be10-ed31d81163a2" (UID: "c50f3d11-a17a-4e29-be10-ed31d81163a2"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.288439 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-run" (OuterVolumeSpecName: "var-run") pod "c50f3d11-a17a-4e29-be10-ed31d81163a2" (UID: "c50f3d11-a17a-4e29-be10-ed31d81163a2"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.289619 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c50f3d11-a17a-4e29-be10-ed31d81163a2-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "c50f3d11-a17a-4e29-be10-ed31d81163a2" (UID: "c50f3d11-a17a-4e29-be10-ed31d81163a2"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.289840 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c50f3d11-a17a-4e29-be10-ed31d81163a2-scripts" (OuterVolumeSpecName: "scripts") pod "c50f3d11-a17a-4e29-be10-ed31d81163a2" (UID: "c50f3d11-a17a-4e29-be10-ed31d81163a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.296450 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c50f3d11-a17a-4e29-be10-ed31d81163a2-kube-api-access-vh7s7" (OuterVolumeSpecName: "kube-api-access-vh7s7") pod "c50f3d11-a17a-4e29-be10-ed31d81163a2" (UID: "c50f3d11-a17a-4e29-be10-ed31d81163a2"). InnerVolumeSpecName "kube-api-access-vh7s7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.390605 4814 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.390654 4814 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.390672 4814 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c50f3d11-a17a-4e29-be10-ed31d81163a2-var-run\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.390689 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c50f3d11-a17a-4e29-be10-ed31d81163a2-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.390706 4814 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c50f3d11-a17a-4e29-be10-ed31d81163a2-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.390730 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vh7s7\" (UniqueName: \"kubernetes.io/projected/c50f3d11-a17a-4e29-be10-ed31d81163a2-kube-api-access-vh7s7\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.642966 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9tkjl-config-l7jb5" event={"ID":"c50f3d11-a17a-4e29-be10-ed31d81163a2","Type":"ContainerDied","Data":"1c039d0fce03df46f36646593980ae82e9385c0071576ae497bd0fbc84a54a58"} Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.643021 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c039d0fce03df46f36646593980ae82e9385c0071576ae497bd0fbc84a54a58" Feb 27 16:42:53 crc kubenswrapper[4814]: I0227 16:42:53.643019 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9tkjl-config-l7jb5" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.275174 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-9tkjl-config-l7jb5"] Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.282496 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-9tkjl-config-l7jb5"] Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.401476 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-9tkjl-config-5fr72"] Feb 27 16:42:54 crc kubenswrapper[4814]: E0227 16:42:54.401886 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c50f3d11-a17a-4e29-be10-ed31d81163a2" containerName="ovn-config" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.401898 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="c50f3d11-a17a-4e29-be10-ed31d81163a2" containerName="ovn-config" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.402030 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="c50f3d11-a17a-4e29-be10-ed31d81163a2" containerName="ovn-config" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.402531 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.408905 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.420396 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9tkjl-config-5fr72"] Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.498388 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c50f3d11-a17a-4e29-be10-ed31d81163a2" path="/var/lib/kubelet/pods/c50f3d11-a17a-4e29-be10-ed31d81163a2/volumes" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.508082 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqbqs\" (UniqueName: \"kubernetes.io/projected/0221b9ca-7350-404c-b754-426f9d900a88-kube-api-access-gqbqs\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.508153 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-run-ovn\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.508189 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0221b9ca-7350-404c-b754-426f9d900a88-additional-scripts\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.508206 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-log-ovn\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.508240 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-run\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.508276 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0221b9ca-7350-404c-b754-426f9d900a88-scripts\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.610001 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-run-ovn\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.610348 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0221b9ca-7350-404c-b754-426f9d900a88-additional-scripts\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.610343 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-run-ovn\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.610367 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-log-ovn\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.610554 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-run\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.610637 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0221b9ca-7350-404c-b754-426f9d900a88-scripts\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.610722 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-run\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.610895 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqbqs\" (UniqueName: \"kubernetes.io/projected/0221b9ca-7350-404c-b754-426f9d900a88-kube-api-access-gqbqs\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.611097 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0221b9ca-7350-404c-b754-426f9d900a88-additional-scripts\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.611192 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-log-ovn\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.612574 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0221b9ca-7350-404c-b754-426f9d900a88-scripts\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.626306 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqbqs\" (UniqueName: \"kubernetes.io/projected/0221b9ca-7350-404c-b754-426f9d900a88-kube-api-access-gqbqs\") pod \"ovn-controller-9tkjl-config-5fr72\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:54 crc kubenswrapper[4814]: I0227 16:42:54.863537 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:55 crc kubenswrapper[4814]: I0227 16:42:55.383648 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9tkjl-config-5fr72"] Feb 27 16:42:55 crc kubenswrapper[4814]: I0227 16:42:55.666855 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9746d8ab-e2c8-419c-8327-6705260b31dd","Type":"ContainerStarted","Data":"3d0dc59a7eb6f7f153bb528b2891b96d948043c5dbb7b2380b741409395895ac"} Feb 27 16:42:55 crc kubenswrapper[4814]: I0227 16:42:55.667357 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9746d8ab-e2c8-419c-8327-6705260b31dd","Type":"ContainerStarted","Data":"7c5dd9f784bedb411f2721e503658e22d20d5a3657131d07b6a83e9f9c8a1b59"} Feb 27 16:42:55 crc kubenswrapper[4814]: I0227 16:42:55.667384 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9746d8ab-e2c8-419c-8327-6705260b31dd","Type":"ContainerStarted","Data":"12928a0f4e16ccaaded529db7dd791a19b9b9cd8b381ef03744c586db6a96d8f"} Feb 27 16:42:55 crc kubenswrapper[4814]: I0227 16:42:55.667407 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9746d8ab-e2c8-419c-8327-6705260b31dd","Type":"ContainerStarted","Data":"21d04d119e152569ea7e0e61d6f8d2181f3374c9de7e92702a73c63e45f1a42c"} Feb 27 16:42:55 crc kubenswrapper[4814]: I0227 16:42:55.669083 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-k85vk" event={"ID":"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9","Type":"ContainerStarted","Data":"002ed9a3bed92d76b420850d31727c7776c5ef76ad9fe80068395f30a52215b5"} Feb 27 16:42:55 crc kubenswrapper[4814]: I0227 16:42:55.670088 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9tkjl-config-5fr72" event={"ID":"0221b9ca-7350-404c-b754-426f9d900a88","Type":"ContainerStarted","Data":"b9b1a7255a3edcd6f155982eaa80d460edac471ee3ddfee1c5d811ffbb9061ca"} Feb 27 16:42:55 crc kubenswrapper[4814]: I0227 16:42:55.683927 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-k85vk" podStartSLOduration=2.692240724 podStartE2EDuration="16.683904107s" podCreationTimestamp="2026-02-27 16:42:39 +0000 UTC" firstStartedPulling="2026-02-27 16:42:40.36266635 +0000 UTC m=+1172.815291190" lastFinishedPulling="2026-02-27 16:42:54.354329743 +0000 UTC m=+1186.806954573" observedRunningTime="2026-02-27 16:42:55.680205501 +0000 UTC m=+1188.132830341" watchObservedRunningTime="2026-02-27 16:42:55.683904107 +0000 UTC m=+1188.136528937" Feb 27 16:42:56 crc kubenswrapper[4814]: I0227 16:42:56.006787 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="8bfc5fb5-c9ab-4108-915b-e6669b332e05" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Feb 27 16:42:56 crc kubenswrapper[4814]: I0227 16:42:56.346407 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="58660097-6b23-4016-98b4-6a10978c0887" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Feb 27 16:42:56 crc kubenswrapper[4814]: I0227 16:42:56.699499 4814 generic.go:334] "Generic (PLEG): container finished" podID="0221b9ca-7350-404c-b754-426f9d900a88" containerID="49a68508357e578f6c9b99dce6fa55364fe996ac9b6d9f43c57b556c3b745576" exitCode=0 Feb 27 16:42:56 crc kubenswrapper[4814]: I0227 16:42:56.699651 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9tkjl-config-5fr72" event={"ID":"0221b9ca-7350-404c-b754-426f9d900a88","Type":"ContainerDied","Data":"49a68508357e578f6c9b99dce6fa55364fe996ac9b6d9f43c57b556c3b745576"} Feb 27 16:42:56 crc kubenswrapper[4814]: I0227 16:42:56.712442 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9746d8ab-e2c8-419c-8327-6705260b31dd","Type":"ContainerStarted","Data":"b08aa5bf9aa77ad8160cb120a8edc9e7f72c5244c8d94500a0f6efe829c58a57"} Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.097611 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.274221 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-log-ovn\") pod \"0221b9ca-7350-404c-b754-426f9d900a88\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.274374 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqbqs\" (UniqueName: \"kubernetes.io/projected/0221b9ca-7350-404c-b754-426f9d900a88-kube-api-access-gqbqs\") pod \"0221b9ca-7350-404c-b754-426f9d900a88\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.274459 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-run\") pod \"0221b9ca-7350-404c-b754-426f9d900a88\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.274497 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0221b9ca-7350-404c-b754-426f9d900a88-scripts\") pod \"0221b9ca-7350-404c-b754-426f9d900a88\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.274654 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0221b9ca-7350-404c-b754-426f9d900a88-additional-scripts\") pod \"0221b9ca-7350-404c-b754-426f9d900a88\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.274716 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-run-ovn\") pod \"0221b9ca-7350-404c-b754-426f9d900a88\" (UID: \"0221b9ca-7350-404c-b754-426f9d900a88\") " Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.275203 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "0221b9ca-7350-404c-b754-426f9d900a88" (UID: "0221b9ca-7350-404c-b754-426f9d900a88"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.275352 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "0221b9ca-7350-404c-b754-426f9d900a88" (UID: "0221b9ca-7350-404c-b754-426f9d900a88"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.275424 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-run" (OuterVolumeSpecName: "var-run") pod "0221b9ca-7350-404c-b754-426f9d900a88" (UID: "0221b9ca-7350-404c-b754-426f9d900a88"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.276023 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0221b9ca-7350-404c-b754-426f9d900a88-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "0221b9ca-7350-404c-b754-426f9d900a88" (UID: "0221b9ca-7350-404c-b754-426f9d900a88"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.276082 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0221b9ca-7350-404c-b754-426f9d900a88-scripts" (OuterVolumeSpecName: "scripts") pod "0221b9ca-7350-404c-b754-426f9d900a88" (UID: "0221b9ca-7350-404c-b754-426f9d900a88"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.283406 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0221b9ca-7350-404c-b754-426f9d900a88-kube-api-access-gqbqs" (OuterVolumeSpecName: "kube-api-access-gqbqs") pod "0221b9ca-7350-404c-b754-426f9d900a88" (UID: "0221b9ca-7350-404c-b754-426f9d900a88"). InnerVolumeSpecName "kube-api-access-gqbqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.377462 4814 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0221b9ca-7350-404c-b754-426f9d900a88-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.377504 4814 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.377515 4814 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.377525 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqbqs\" (UniqueName: \"kubernetes.io/projected/0221b9ca-7350-404c-b754-426f9d900a88-kube-api-access-gqbqs\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.377540 4814 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0221b9ca-7350-404c-b754-426f9d900a88-var-run\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.377550 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0221b9ca-7350-404c-b754-426f9d900a88-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.732378 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9tkjl-config-5fr72" event={"ID":"0221b9ca-7350-404c-b754-426f9d900a88","Type":"ContainerDied","Data":"b9b1a7255a3edcd6f155982eaa80d460edac471ee3ddfee1c5d811ffbb9061ca"} Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.732414 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9tkjl-config-5fr72" Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.732420 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9b1a7255a3edcd6f155982eaa80d460edac471ee3ddfee1c5d811ffbb9061ca" Feb 27 16:42:58 crc kubenswrapper[4814]: I0227 16:42:58.748959 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9746d8ab-e2c8-419c-8327-6705260b31dd","Type":"ContainerStarted","Data":"a5e0c1c2a32a3684b9a4e2855b3c43bba3a96f3a70abf10d5974c31c2f83b766"} Feb 27 16:42:59 crc kubenswrapper[4814]: I0227 16:42:59.224137 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-9tkjl-config-5fr72"] Feb 27 16:42:59 crc kubenswrapper[4814]: I0227 16:42:59.230669 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-9tkjl-config-5fr72"] Feb 27 16:42:59 crc kubenswrapper[4814]: I0227 16:42:59.772321 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9746d8ab-e2c8-419c-8327-6705260b31dd","Type":"ContainerStarted","Data":"df5b18b4f8e996746f41f0df5fc24bb820b6b87acb2d6460abacb7a762901c34"} Feb 27 16:42:59 crc kubenswrapper[4814]: I0227 16:42:59.846623 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=24.243121129 podStartE2EDuration="38.846598578s" podCreationTimestamp="2026-02-27 16:42:21 +0000 UTC" firstStartedPulling="2026-02-27 16:42:39.758760722 +0000 UTC m=+1172.211385582" lastFinishedPulling="2026-02-27 16:42:54.362238161 +0000 UTC m=+1186.814863031" observedRunningTime="2026-02-27 16:42:59.826475116 +0000 UTC m=+1192.279099976" watchObservedRunningTime="2026-02-27 16:42:59.846598578 +0000 UTC m=+1192.299223438" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.138355 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-wz2lv"] Feb 27 16:43:00 crc kubenswrapper[4814]: E0227 16:43:00.138665 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0221b9ca-7350-404c-b754-426f9d900a88" containerName="ovn-config" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.138681 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="0221b9ca-7350-404c-b754-426f9d900a88" containerName="ovn-config" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.138836 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="0221b9ca-7350-404c-b754-426f9d900a88" containerName="ovn-config" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.139575 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.145757 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.152631 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-wz2lv"] Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.310199 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.310376 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.310416 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.310512 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxdt4\" (UniqueName: \"kubernetes.io/projected/4ae44728-ad5a-4167-8823-b1b273265a29-kube-api-access-hxdt4\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.310543 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.310646 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-config\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.411979 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxdt4\" (UniqueName: \"kubernetes.io/projected/4ae44728-ad5a-4167-8823-b1b273265a29-kube-api-access-hxdt4\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.412336 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.413115 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.413731 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-config\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.413792 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-config\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.413835 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.413910 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.413957 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.414515 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.415362 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.415865 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.440822 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxdt4\" (UniqueName: \"kubernetes.io/projected/4ae44728-ad5a-4167-8823-b1b273265a29-kube-api-access-hxdt4\") pod \"dnsmasq-dns-77585f5f8c-wz2lv\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.456158 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.524166 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0221b9ca-7350-404c-b754-426f9d900a88" path="/var/lib/kubelet/pods/0221b9ca-7350-404c-b754-426f9d900a88/volumes" Feb 27 16:43:00 crc kubenswrapper[4814]: I0227 16:43:00.937981 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-wz2lv"] Feb 27 16:43:00 crc kubenswrapper[4814]: W0227 16:43:00.946536 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ae44728_ad5a_4167_8823_b1b273265a29.slice/crio-2bde3f60b42237111ff14da9a2d03f473f48f717185b237dbc39aa8cdbf0949b WatchSource:0}: Error finding container 2bde3f60b42237111ff14da9a2d03f473f48f717185b237dbc39aa8cdbf0949b: Status 404 returned error can't find the container with id 2bde3f60b42237111ff14da9a2d03f473f48f717185b237dbc39aa8cdbf0949b Feb 27 16:43:01 crc kubenswrapper[4814]: I0227 16:43:01.791090 4814 generic.go:334] "Generic (PLEG): container finished" podID="4ae44728-ad5a-4167-8823-b1b273265a29" containerID="0f868ed4357392cf34291c9a8d69618903599c078e3052e5c429cb032301b99b" exitCode=0 Feb 27 16:43:01 crc kubenswrapper[4814]: I0227 16:43:01.791760 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" event={"ID":"4ae44728-ad5a-4167-8823-b1b273265a29","Type":"ContainerDied","Data":"0f868ed4357392cf34291c9a8d69618903599c078e3052e5c429cb032301b99b"} Feb 27 16:43:01 crc kubenswrapper[4814]: I0227 16:43:01.791925 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" event={"ID":"4ae44728-ad5a-4167-8823-b1b273265a29","Type":"ContainerStarted","Data":"2bde3f60b42237111ff14da9a2d03f473f48f717185b237dbc39aa8cdbf0949b"} Feb 27 16:43:02 crc kubenswrapper[4814]: I0227 16:43:02.802713 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" event={"ID":"4ae44728-ad5a-4167-8823-b1b273265a29","Type":"ContainerStarted","Data":"eb9a34f2ce8980b3b94acf9e587c2bde662f1e06e113cc61530af63123fca3fa"} Feb 27 16:43:02 crc kubenswrapper[4814]: I0227 16:43:02.803162 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:02 crc kubenswrapper[4814]: I0227 16:43:02.837103 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" podStartSLOduration=2.837077548 podStartE2EDuration="2.837077548s" podCreationTimestamp="2026-02-27 16:43:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:43:02.832336438 +0000 UTC m=+1195.284961308" watchObservedRunningTime="2026-02-27 16:43:02.837077548 +0000 UTC m=+1195.289702418" Feb 27 16:43:03 crc kubenswrapper[4814]: I0227 16:43:03.816129 4814 generic.go:334] "Generic (PLEG): container finished" podID="7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9" containerID="002ed9a3bed92d76b420850d31727c7776c5ef76ad9fe80068395f30a52215b5" exitCode=0 Feb 27 16:43:03 crc kubenswrapper[4814]: I0227 16:43:03.816245 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-k85vk" event={"ID":"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9","Type":"ContainerDied","Data":"002ed9a3bed92d76b420850d31727c7776c5ef76ad9fe80068395f30a52215b5"} Feb 27 16:43:05 crc kubenswrapper[4814]: I0227 16:43:05.367911 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-k85vk" Feb 27 16:43:05 crc kubenswrapper[4814]: I0227 16:43:05.523627 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cnfr\" (UniqueName: \"kubernetes.io/projected/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-kube-api-access-7cnfr\") pod \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\" (UID: \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\") " Feb 27 16:43:05 crc kubenswrapper[4814]: I0227 16:43:05.523694 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-combined-ca-bundle\") pod \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\" (UID: \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\") " Feb 27 16:43:05 crc kubenswrapper[4814]: I0227 16:43:05.523741 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-config-data\") pod \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\" (UID: \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\") " Feb 27 16:43:05 crc kubenswrapper[4814]: I0227 16:43:05.524453 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-db-sync-config-data\") pod \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\" (UID: \"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9\") " Feb 27 16:43:05 crc kubenswrapper[4814]: I0227 16:43:05.532576 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9" (UID: "7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:05 crc kubenswrapper[4814]: I0227 16:43:05.532787 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-kube-api-access-7cnfr" (OuterVolumeSpecName: "kube-api-access-7cnfr") pod "7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9" (UID: "7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9"). InnerVolumeSpecName "kube-api-access-7cnfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:05 crc kubenswrapper[4814]: I0227 16:43:05.579382 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9" (UID: "7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:05 crc kubenswrapper[4814]: I0227 16:43:05.606812 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-config-data" (OuterVolumeSpecName: "config-data") pod "7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9" (UID: "7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:05 crc kubenswrapper[4814]: I0227 16:43:05.626028 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cnfr\" (UniqueName: \"kubernetes.io/projected/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-kube-api-access-7cnfr\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:05 crc kubenswrapper[4814]: I0227 16:43:05.626064 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:05 crc kubenswrapper[4814]: I0227 16:43:05.626078 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:05 crc kubenswrapper[4814]: I0227 16:43:05.626092 4814 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:05 crc kubenswrapper[4814]: I0227 16:43:05.846157 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-k85vk" event={"ID":"7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9","Type":"ContainerDied","Data":"5a843f21106d4575e4dcea76ea8f25fea010552528e7252f3649804690d91137"} Feb 27 16:43:05 crc kubenswrapper[4814]: I0227 16:43:05.846216 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a843f21106d4575e4dcea76ea8f25fea010552528e7252f3649804690d91137" Feb 27 16:43:05 crc kubenswrapper[4814]: I0227 16:43:05.846424 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-k85vk" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.008635 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.347458 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.390672 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-wz2lv"] Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.390882 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" podUID="4ae44728-ad5a-4167-8823-b1b273265a29" containerName="dnsmasq-dns" containerID="cri-o://eb9a34f2ce8980b3b94acf9e587c2bde662f1e06e113cc61530af63123fca3fa" gracePeriod=10 Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.447689 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-qdhj8"] Feb 27 16:43:06 crc kubenswrapper[4814]: E0227 16:43:06.448023 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9" containerName="glance-db-sync" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.448039 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9" containerName="glance-db-sync" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.448202 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9" containerName="glance-db-sync" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.448982 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.459846 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-qdhj8"] Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.541631 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.541682 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.541760 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.541798 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp558\" (UniqueName: \"kubernetes.io/projected/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-kube-api-access-hp558\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.541841 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.541894 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-config\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.643407 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.643472 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp558\" (UniqueName: \"kubernetes.io/projected/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-kube-api-access-hp558\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.643516 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.643543 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-config\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.643583 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.643598 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.645403 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.646114 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-config\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.646396 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.646434 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.646516 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.668622 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp558\" (UniqueName: \"kubernetes.io/projected/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-kube-api-access-hp558\") pod \"dnsmasq-dns-7ff5475cc9-qdhj8\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.786061 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.856104 4814 generic.go:334] "Generic (PLEG): container finished" podID="4ae44728-ad5a-4167-8823-b1b273265a29" containerID="eb9a34f2ce8980b3b94acf9e587c2bde662f1e06e113cc61530af63123fca3fa" exitCode=0 Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.856144 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" event={"ID":"4ae44728-ad5a-4167-8823-b1b273265a29","Type":"ContainerDied","Data":"eb9a34f2ce8980b3b94acf9e587c2bde662f1e06e113cc61530af63123fca3fa"} Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.856169 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" event={"ID":"4ae44728-ad5a-4167-8823-b1b273265a29","Type":"ContainerDied","Data":"2bde3f60b42237111ff14da9a2d03f473f48f717185b237dbc39aa8cdbf0949b"} Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.856181 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bde3f60b42237111ff14da9a2d03f473f48f717185b237dbc39aa8cdbf0949b" Feb 27 16:43:06 crc kubenswrapper[4814]: I0227 16:43:06.861205 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:06.948909 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxdt4\" (UniqueName: \"kubernetes.io/projected/4ae44728-ad5a-4167-8823-b1b273265a29-kube-api-access-hxdt4\") pod \"4ae44728-ad5a-4167-8823-b1b273265a29\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:06.949186 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-ovsdbserver-sb\") pod \"4ae44728-ad5a-4167-8823-b1b273265a29\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:06.949265 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-ovsdbserver-nb\") pod \"4ae44728-ad5a-4167-8823-b1b273265a29\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:06.949313 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-dns-svc\") pod \"4ae44728-ad5a-4167-8823-b1b273265a29\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:06.949345 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-config\") pod \"4ae44728-ad5a-4167-8823-b1b273265a29\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:06.949399 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-dns-swift-storage-0\") pod \"4ae44728-ad5a-4167-8823-b1b273265a29\" (UID: \"4ae44728-ad5a-4167-8823-b1b273265a29\") " Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:06.970455 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ae44728-ad5a-4167-8823-b1b273265a29-kube-api-access-hxdt4" (OuterVolumeSpecName: "kube-api-access-hxdt4") pod "4ae44728-ad5a-4167-8823-b1b273265a29" (UID: "4ae44728-ad5a-4167-8823-b1b273265a29"). InnerVolumeSpecName "kube-api-access-hxdt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:06.988756 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4ae44728-ad5a-4167-8823-b1b273265a29" (UID: "4ae44728-ad5a-4167-8823-b1b273265a29"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:06.990038 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-config" (OuterVolumeSpecName: "config") pod "4ae44728-ad5a-4167-8823-b1b273265a29" (UID: "4ae44728-ad5a-4167-8823-b1b273265a29"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:06.991808 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4ae44728-ad5a-4167-8823-b1b273265a29" (UID: "4ae44728-ad5a-4167-8823-b1b273265a29"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:06.995680 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4ae44728-ad5a-4167-8823-b1b273265a29" (UID: "4ae44728-ad5a-4167-8823-b1b273265a29"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:06.996502 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4ae44728-ad5a-4167-8823-b1b273265a29" (UID: "4ae44728-ad5a-4167-8823-b1b273265a29"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:07.051033 4814 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:07.051068 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:07.051080 4814 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:07.051091 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxdt4\" (UniqueName: \"kubernetes.io/projected/4ae44728-ad5a-4167-8823-b1b273265a29-kube-api-access-hxdt4\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:07.051102 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:07.051114 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ae44728-ad5a-4167-8823-b1b273265a29-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:07.876547 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-wz2lv" Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:07.945084 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-wz2lv"] Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:07.957941 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-wz2lv"] Feb 27 16:43:07 crc kubenswrapper[4814]: I0227 16:43:07.967989 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-qdhj8"] Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.141098 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-tvmcg"] Feb 27 16:43:08 crc kubenswrapper[4814]: E0227 16:43:08.148317 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ae44728-ad5a-4167-8823-b1b273265a29" containerName="dnsmasq-dns" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.148410 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ae44728-ad5a-4167-8823-b1b273265a29" containerName="dnsmasq-dns" Feb 27 16:43:08 crc kubenswrapper[4814]: E0227 16:43:08.148470 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ae44728-ad5a-4167-8823-b1b273265a29" containerName="init" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.148514 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ae44728-ad5a-4167-8823-b1b273265a29" containerName="init" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.148727 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ae44728-ad5a-4167-8823-b1b273265a29" containerName="dnsmasq-dns" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.149401 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tvmcg" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.158791 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-tvmcg"] Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.253351 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-2e37-account-create-update-dxdg7"] Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.259023 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2e37-account-create-update-dxdg7" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.265998 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.270345 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxszf\" (UniqueName: \"kubernetes.io/projected/b885ee2d-5c14-40b1-83c5-068d2bb0ef96-kube-api-access-pxszf\") pod \"cinder-db-create-tvmcg\" (UID: \"b885ee2d-5c14-40b1-83c5-068d2bb0ef96\") " pod="openstack/cinder-db-create-tvmcg" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.270444 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b885ee2d-5c14-40b1-83c5-068d2bb0ef96-operator-scripts\") pod \"cinder-db-create-tvmcg\" (UID: \"b885ee2d-5c14-40b1-83c5-068d2bb0ef96\") " pod="openstack/cinder-db-create-tvmcg" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.271898 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-2e37-account-create-update-dxdg7"] Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.337781 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-vphqv"] Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.338740 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vphqv" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.356080 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-6479-account-create-update-5jbnw"] Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.357096 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6479-account-create-update-5jbnw" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.360142 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.370007 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-vphqv"] Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.373373 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxszf\" (UniqueName: \"kubernetes.io/projected/b885ee2d-5c14-40b1-83c5-068d2bb0ef96-kube-api-access-pxszf\") pod \"cinder-db-create-tvmcg\" (UID: \"b885ee2d-5c14-40b1-83c5-068d2bb0ef96\") " pod="openstack/cinder-db-create-tvmcg" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.373442 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s5g8\" (UniqueName: \"kubernetes.io/projected/bd1f40c7-a082-4356-848b-578cb58f1015-kube-api-access-8s5g8\") pod \"cinder-2e37-account-create-update-dxdg7\" (UID: \"bd1f40c7-a082-4356-848b-578cb58f1015\") " pod="openstack/cinder-2e37-account-create-update-dxdg7" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.373499 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd1f40c7-a082-4356-848b-578cb58f1015-operator-scripts\") pod \"cinder-2e37-account-create-update-dxdg7\" (UID: \"bd1f40c7-a082-4356-848b-578cb58f1015\") " pod="openstack/cinder-2e37-account-create-update-dxdg7" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.373553 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b885ee2d-5c14-40b1-83c5-068d2bb0ef96-operator-scripts\") pod \"cinder-db-create-tvmcg\" (UID: \"b885ee2d-5c14-40b1-83c5-068d2bb0ef96\") " pod="openstack/cinder-db-create-tvmcg" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.374201 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b885ee2d-5c14-40b1-83c5-068d2bb0ef96-operator-scripts\") pod \"cinder-db-create-tvmcg\" (UID: \"b885ee2d-5c14-40b1-83c5-068d2bb0ef96\") " pod="openstack/cinder-db-create-tvmcg" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.377915 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6479-account-create-update-5jbnw"] Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.391273 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxszf\" (UniqueName: \"kubernetes.io/projected/b885ee2d-5c14-40b1-83c5-068d2bb0ef96-kube-api-access-pxszf\") pod \"cinder-db-create-tvmcg\" (UID: \"b885ee2d-5c14-40b1-83c5-068d2bb0ef96\") " pod="openstack/cinder-db-create-tvmcg" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.422910 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-l4sht"] Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.423811 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-l4sht" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.427393 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.427615 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.427831 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.429245 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-m5bsz" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.439697 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-l4sht"] Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.474927 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9-operator-scripts\") pod \"barbican-db-create-vphqv\" (UID: \"0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9\") " pod="openstack/barbican-db-create-vphqv" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.474995 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f-operator-scripts\") pod \"barbican-6479-account-create-update-5jbnw\" (UID: \"0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f\") " pod="openstack/barbican-6479-account-create-update-5jbnw" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.475027 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pkj7\" (UniqueName: \"kubernetes.io/projected/0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f-kube-api-access-8pkj7\") pod \"barbican-6479-account-create-update-5jbnw\" (UID: \"0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f\") " pod="openstack/barbican-6479-account-create-update-5jbnw" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.475092 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s5g8\" (UniqueName: \"kubernetes.io/projected/bd1f40c7-a082-4356-848b-578cb58f1015-kube-api-access-8s5g8\") pod \"cinder-2e37-account-create-update-dxdg7\" (UID: \"bd1f40c7-a082-4356-848b-578cb58f1015\") " pod="openstack/cinder-2e37-account-create-update-dxdg7" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.475124 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd1f40c7-a082-4356-848b-578cb58f1015-operator-scripts\") pod \"cinder-2e37-account-create-update-dxdg7\" (UID: \"bd1f40c7-a082-4356-848b-578cb58f1015\") " pod="openstack/cinder-2e37-account-create-update-dxdg7" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.475190 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p422l\" (UniqueName: \"kubernetes.io/projected/0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9-kube-api-access-p422l\") pod \"barbican-db-create-vphqv\" (UID: \"0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9\") " pod="openstack/barbican-db-create-vphqv" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.476289 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd1f40c7-a082-4356-848b-578cb58f1015-operator-scripts\") pod \"cinder-2e37-account-create-update-dxdg7\" (UID: \"bd1f40c7-a082-4356-848b-578cb58f1015\") " pod="openstack/cinder-2e37-account-create-update-dxdg7" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.482119 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-lgxt2"] Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.482370 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tvmcg" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.483058 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lgxt2" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.501772 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s5g8\" (UniqueName: \"kubernetes.io/projected/bd1f40c7-a082-4356-848b-578cb58f1015-kube-api-access-8s5g8\") pod \"cinder-2e37-account-create-update-dxdg7\" (UID: \"bd1f40c7-a082-4356-848b-578cb58f1015\") " pod="openstack/cinder-2e37-account-create-update-dxdg7" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.572203 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ae44728-ad5a-4167-8823-b1b273265a29" path="/var/lib/kubelet/pods/4ae44728-ad5a-4167-8823-b1b273265a29/volumes" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.572796 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-lgxt2"] Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.573955 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2e37-account-create-update-dxdg7" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.576333 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6779899f-08d7-4de4-884c-070af4597386-config-data\") pod \"keystone-db-sync-l4sht\" (UID: \"6779899f-08d7-4de4-884c-070af4597386\") " pod="openstack/keystone-db-sync-l4sht" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.576366 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6779899f-08d7-4de4-884c-070af4597386-combined-ca-bundle\") pod \"keystone-db-sync-l4sht\" (UID: \"6779899f-08d7-4de4-884c-070af4597386\") " pod="openstack/keystone-db-sync-l4sht" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.576398 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4q2c\" (UniqueName: \"kubernetes.io/projected/6779899f-08d7-4de4-884c-070af4597386-kube-api-access-t4q2c\") pod \"keystone-db-sync-l4sht\" (UID: \"6779899f-08d7-4de4-884c-070af4597386\") " pod="openstack/keystone-db-sync-l4sht" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.576438 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p422l\" (UniqueName: \"kubernetes.io/projected/0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9-kube-api-access-p422l\") pod \"barbican-db-create-vphqv\" (UID: \"0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9\") " pod="openstack/barbican-db-create-vphqv" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.576494 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9-operator-scripts\") pod \"barbican-db-create-vphqv\" (UID: \"0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9\") " pod="openstack/barbican-db-create-vphqv" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.576528 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f-operator-scripts\") pod \"barbican-6479-account-create-update-5jbnw\" (UID: \"0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f\") " pod="openstack/barbican-6479-account-create-update-5jbnw" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.576620 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pkj7\" (UniqueName: \"kubernetes.io/projected/0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f-kube-api-access-8pkj7\") pod \"barbican-6479-account-create-update-5jbnw\" (UID: \"0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f\") " pod="openstack/barbican-6479-account-create-update-5jbnw" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.576675 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c2ab590-4b44-4d31-9324-2abffd328742-operator-scripts\") pod \"neutron-db-create-lgxt2\" (UID: \"9c2ab590-4b44-4d31-9324-2abffd328742\") " pod="openstack/neutron-db-create-lgxt2" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.576761 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffwbn\" (UniqueName: \"kubernetes.io/projected/9c2ab590-4b44-4d31-9324-2abffd328742-kube-api-access-ffwbn\") pod \"neutron-db-create-lgxt2\" (UID: \"9c2ab590-4b44-4d31-9324-2abffd328742\") " pod="openstack/neutron-db-create-lgxt2" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.578441 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9-operator-scripts\") pod \"barbican-db-create-vphqv\" (UID: \"0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9\") " pod="openstack/barbican-db-create-vphqv" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.578536 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f-operator-scripts\") pod \"barbican-6479-account-create-update-5jbnw\" (UID: \"0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f\") " pod="openstack/barbican-6479-account-create-update-5jbnw" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.593645 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pkj7\" (UniqueName: \"kubernetes.io/projected/0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f-kube-api-access-8pkj7\") pod \"barbican-6479-account-create-update-5jbnw\" (UID: \"0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f\") " pod="openstack/barbican-6479-account-create-update-5jbnw" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.624766 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p422l\" (UniqueName: \"kubernetes.io/projected/0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9-kube-api-access-p422l\") pod \"barbican-db-create-vphqv\" (UID: \"0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9\") " pod="openstack/barbican-db-create-vphqv" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.652442 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-2845-account-create-update-pzlvv"] Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.660116 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vphqv" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.667917 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2845-account-create-update-pzlvv"] Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.667999 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2845-account-create-update-pzlvv" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.670614 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.672987 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6479-account-create-update-5jbnw" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.688994 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c2ab590-4b44-4d31-9324-2abffd328742-operator-scripts\") pod \"neutron-db-create-lgxt2\" (UID: \"9c2ab590-4b44-4d31-9324-2abffd328742\") " pod="openstack/neutron-db-create-lgxt2" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.689057 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffwbn\" (UniqueName: \"kubernetes.io/projected/9c2ab590-4b44-4d31-9324-2abffd328742-kube-api-access-ffwbn\") pod \"neutron-db-create-lgxt2\" (UID: \"9c2ab590-4b44-4d31-9324-2abffd328742\") " pod="openstack/neutron-db-create-lgxt2" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.689102 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6779899f-08d7-4de4-884c-070af4597386-config-data\") pod \"keystone-db-sync-l4sht\" (UID: \"6779899f-08d7-4de4-884c-070af4597386\") " pod="openstack/keystone-db-sync-l4sht" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.689122 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6779899f-08d7-4de4-884c-070af4597386-combined-ca-bundle\") pod \"keystone-db-sync-l4sht\" (UID: \"6779899f-08d7-4de4-884c-070af4597386\") " pod="openstack/keystone-db-sync-l4sht" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.689144 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4q2c\" (UniqueName: \"kubernetes.io/projected/6779899f-08d7-4de4-884c-070af4597386-kube-api-access-t4q2c\") pod \"keystone-db-sync-l4sht\" (UID: \"6779899f-08d7-4de4-884c-070af4597386\") " pod="openstack/keystone-db-sync-l4sht" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.689988 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c2ab590-4b44-4d31-9324-2abffd328742-operator-scripts\") pod \"neutron-db-create-lgxt2\" (UID: \"9c2ab590-4b44-4d31-9324-2abffd328742\") " pod="openstack/neutron-db-create-lgxt2" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.693349 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6779899f-08d7-4de4-884c-070af4597386-config-data\") pod \"keystone-db-sync-l4sht\" (UID: \"6779899f-08d7-4de4-884c-070af4597386\") " pod="openstack/keystone-db-sync-l4sht" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.698378 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6779899f-08d7-4de4-884c-070af4597386-combined-ca-bundle\") pod \"keystone-db-sync-l4sht\" (UID: \"6779899f-08d7-4de4-884c-070af4597386\") " pod="openstack/keystone-db-sync-l4sht" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.711004 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffwbn\" (UniqueName: \"kubernetes.io/projected/9c2ab590-4b44-4d31-9324-2abffd328742-kube-api-access-ffwbn\") pod \"neutron-db-create-lgxt2\" (UID: \"9c2ab590-4b44-4d31-9324-2abffd328742\") " pod="openstack/neutron-db-create-lgxt2" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.724888 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4q2c\" (UniqueName: \"kubernetes.io/projected/6779899f-08d7-4de4-884c-070af4597386-kube-api-access-t4q2c\") pod \"keystone-db-sync-l4sht\" (UID: \"6779899f-08d7-4de4-884c-070af4597386\") " pod="openstack/keystone-db-sync-l4sht" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.739443 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-l4sht" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.790373 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvp5z\" (UniqueName: \"kubernetes.io/projected/8d3720d6-e79c-44cc-8780-415f2b139a9f-kube-api-access-xvp5z\") pod \"neutron-2845-account-create-update-pzlvv\" (UID: \"8d3720d6-e79c-44cc-8780-415f2b139a9f\") " pod="openstack/neutron-2845-account-create-update-pzlvv" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.790426 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d3720d6-e79c-44cc-8780-415f2b139a9f-operator-scripts\") pod \"neutron-2845-account-create-update-pzlvv\" (UID: \"8d3720d6-e79c-44cc-8780-415f2b139a9f\") " pod="openstack/neutron-2845-account-create-update-pzlvv" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.876519 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lgxt2" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.885465 4814 generic.go:334] "Generic (PLEG): container finished" podID="6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" containerID="25b206576573c6339f705f0fbee3821f6900b276d6dbecd7dce3be15bc5f87d3" exitCode=0 Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.885582 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" event={"ID":"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd","Type":"ContainerDied","Data":"25b206576573c6339f705f0fbee3821f6900b276d6dbecd7dce3be15bc5f87d3"} Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.885650 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" event={"ID":"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd","Type":"ContainerStarted","Data":"2f7e226474e68967bee396e70bed60a819c2b72b5259d688cd8f548731ab7798"} Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.891600 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvp5z\" (UniqueName: \"kubernetes.io/projected/8d3720d6-e79c-44cc-8780-415f2b139a9f-kube-api-access-xvp5z\") pod \"neutron-2845-account-create-update-pzlvv\" (UID: \"8d3720d6-e79c-44cc-8780-415f2b139a9f\") " pod="openstack/neutron-2845-account-create-update-pzlvv" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.891659 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d3720d6-e79c-44cc-8780-415f2b139a9f-operator-scripts\") pod \"neutron-2845-account-create-update-pzlvv\" (UID: \"8d3720d6-e79c-44cc-8780-415f2b139a9f\") " pod="openstack/neutron-2845-account-create-update-pzlvv" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.892511 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d3720d6-e79c-44cc-8780-415f2b139a9f-operator-scripts\") pod \"neutron-2845-account-create-update-pzlvv\" (UID: \"8d3720d6-e79c-44cc-8780-415f2b139a9f\") " pod="openstack/neutron-2845-account-create-update-pzlvv" Feb 27 16:43:08 crc kubenswrapper[4814]: I0227 16:43:08.916570 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvp5z\" (UniqueName: \"kubernetes.io/projected/8d3720d6-e79c-44cc-8780-415f2b139a9f-kube-api-access-xvp5z\") pod \"neutron-2845-account-create-update-pzlvv\" (UID: \"8d3720d6-e79c-44cc-8780-415f2b139a9f\") " pod="openstack/neutron-2845-account-create-update-pzlvv" Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.010714 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-tvmcg"] Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.031093 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2845-account-create-update-pzlvv" Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.146509 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-2e37-account-create-update-dxdg7"] Feb 27 16:43:09 crc kubenswrapper[4814]: W0227 16:43:09.157622 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd1f40c7_a082_4356_848b_578cb58f1015.slice/crio-65804bb23dea7d31d2b25d27353173fdd535d2a5c13254fbe04b194ee5250fb9 WatchSource:0}: Error finding container 65804bb23dea7d31d2b25d27353173fdd535d2a5c13254fbe04b194ee5250fb9: Status 404 returned error can't find the container with id 65804bb23dea7d31d2b25d27353173fdd535d2a5c13254fbe04b194ee5250fb9 Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.163780 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.235114 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-l4sht"] Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.249674 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6479-account-create-update-5jbnw"] Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.257802 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-vphqv"] Feb 27 16:43:09 crc kubenswrapper[4814]: W0227 16:43:09.265636 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6779899f_08d7_4de4_884c_070af4597386.slice/crio-6a5592faccbb78c9b6af076be3533a9c1ea9fd5baef513911445df24b8b0addf WatchSource:0}: Error finding container 6a5592faccbb78c9b6af076be3533a9c1ea9fd5baef513911445df24b8b0addf: Status 404 returned error can't find the container with id 6a5592faccbb78c9b6af076be3533a9c1ea9fd5baef513911445df24b8b0addf Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.278511 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.306303 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2845-account-create-update-pzlvv"] Feb 27 16:43:09 crc kubenswrapper[4814]: W0227 16:43:09.308751 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d3720d6_e79c_44cc_8780_415f2b139a9f.slice/crio-16cde31a3a4e1294bd73e9bf2081447eeaca3c29926f59aea5ed8b06a508379e WatchSource:0}: Error finding container 16cde31a3a4e1294bd73e9bf2081447eeaca3c29926f59aea5ed8b06a508379e: Status 404 returned error can't find the container with id 16cde31a3a4e1294bd73e9bf2081447eeaca3c29926f59aea5ed8b06a508379e Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.375442 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-lgxt2"] Feb 27 16:43:09 crc kubenswrapper[4814]: W0227 16:43:09.393290 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c2ab590_4b44_4d31_9324_2abffd328742.slice/crio-33f1a9fd25989044afe10558bdf23bdd2bc2ff0fea1df6026eb0d27274d133cf WatchSource:0}: Error finding container 33f1a9fd25989044afe10558bdf23bdd2bc2ff0fea1df6026eb0d27274d133cf: Status 404 returned error can't find the container with id 33f1a9fd25989044afe10558bdf23bdd2bc2ff0fea1df6026eb0d27274d133cf Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.897232 4814 generic.go:334] "Generic (PLEG): container finished" podID="bd1f40c7-a082-4356-848b-578cb58f1015" containerID="07f90fb740624019f8172a6b68f873b900dcfd78787dfc6961f3576d32e43e6d" exitCode=0 Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.897303 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2e37-account-create-update-dxdg7" event={"ID":"bd1f40c7-a082-4356-848b-578cb58f1015","Type":"ContainerDied","Data":"07f90fb740624019f8172a6b68f873b900dcfd78787dfc6961f3576d32e43e6d"} Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.897714 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2e37-account-create-update-dxdg7" event={"ID":"bd1f40c7-a082-4356-848b-578cb58f1015","Type":"ContainerStarted","Data":"65804bb23dea7d31d2b25d27353173fdd535d2a5c13254fbe04b194ee5250fb9"} Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.900236 4814 generic.go:334] "Generic (PLEG): container finished" podID="b885ee2d-5c14-40b1-83c5-068d2bb0ef96" containerID="6e34a5ba5d94a1d6d82632db4751abd111e8ee2716a87987714bd0ad494938e1" exitCode=0 Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.900328 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tvmcg" event={"ID":"b885ee2d-5c14-40b1-83c5-068d2bb0ef96","Type":"ContainerDied","Data":"6e34a5ba5d94a1d6d82632db4751abd111e8ee2716a87987714bd0ad494938e1"} Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.900375 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tvmcg" event={"ID":"b885ee2d-5c14-40b1-83c5-068d2bb0ef96","Type":"ContainerStarted","Data":"f08c620f9738d9d005efb3e0f8785bed538d2be949e2275b3b0d3dbc7f900b23"} Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.907756 4814 generic.go:334] "Generic (PLEG): container finished" podID="0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f" containerID="f3f6819b6255e864f982437f471dfed41ea027a8f03f66d18cea0054832d3c1d" exitCode=0 Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.907842 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6479-account-create-update-5jbnw" event={"ID":"0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f","Type":"ContainerDied","Data":"f3f6819b6255e864f982437f471dfed41ea027a8f03f66d18cea0054832d3c1d"} Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.907869 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6479-account-create-update-5jbnw" event={"ID":"0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f","Type":"ContainerStarted","Data":"dd2544d6c76e4508fc55b06d3ecd383b8b4524287ac4abf4ea2650248ae103bd"} Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.911726 4814 generic.go:334] "Generic (PLEG): container finished" podID="9c2ab590-4b44-4d31-9324-2abffd328742" containerID="7d9f4a94e9218f9a4b247299d02a5ad4403f74bc50a6ef5609679e4d8d88bf3a" exitCode=0 Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.911817 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lgxt2" event={"ID":"9c2ab590-4b44-4d31-9324-2abffd328742","Type":"ContainerDied","Data":"7d9f4a94e9218f9a4b247299d02a5ad4403f74bc50a6ef5609679e4d8d88bf3a"} Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.911848 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lgxt2" event={"ID":"9c2ab590-4b44-4d31-9324-2abffd328742","Type":"ContainerStarted","Data":"33f1a9fd25989044afe10558bdf23bdd2bc2ff0fea1df6026eb0d27274d133cf"} Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.920933 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" event={"ID":"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd","Type":"ContainerStarted","Data":"868dca07dd0ef20b6d4dab49ac197397d1fe8947c8339c8c88f4eaf0da2a6e69"} Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.921098 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.923361 4814 generic.go:334] "Generic (PLEG): container finished" podID="0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9" containerID="db2b3be43822ed9bd4853f6086c1e27e0ca957584c3e2042429501bca8827514" exitCode=0 Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.923397 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vphqv" event={"ID":"0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9","Type":"ContainerDied","Data":"db2b3be43822ed9bd4853f6086c1e27e0ca957584c3e2042429501bca8827514"} Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.923443 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vphqv" event={"ID":"0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9","Type":"ContainerStarted","Data":"9912226f1b612f4d9b536d7b400ccde25f68c67215bb6e3c9397cbe63c6b3573"} Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.924932 4814 generic.go:334] "Generic (PLEG): container finished" podID="8d3720d6-e79c-44cc-8780-415f2b139a9f" containerID="d0ae7868c8f87fcbb95cab422c02698a0c3eac93473dcbe8ef0196aa8e9743eb" exitCode=0 Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.924985 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2845-account-create-update-pzlvv" event={"ID":"8d3720d6-e79c-44cc-8780-415f2b139a9f","Type":"ContainerDied","Data":"d0ae7868c8f87fcbb95cab422c02698a0c3eac93473dcbe8ef0196aa8e9743eb"} Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.925025 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2845-account-create-update-pzlvv" event={"ID":"8d3720d6-e79c-44cc-8780-415f2b139a9f","Type":"ContainerStarted","Data":"16cde31a3a4e1294bd73e9bf2081447eeaca3c29926f59aea5ed8b06a508379e"} Feb 27 16:43:09 crc kubenswrapper[4814]: I0227 16:43:09.926267 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-l4sht" event={"ID":"6779899f-08d7-4de4-884c-070af4597386","Type":"ContainerStarted","Data":"6a5592faccbb78c9b6af076be3533a9c1ea9fd5baef513911445df24b8b0addf"} Feb 27 16:43:10 crc kubenswrapper[4814]: I0227 16:43:10.029318 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" podStartSLOduration=4.029295352 podStartE2EDuration="4.029295352s" podCreationTimestamp="2026-02-27 16:43:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:43:10.006974142 +0000 UTC m=+1202.459598972" watchObservedRunningTime="2026-02-27 16:43:10.029295352 +0000 UTC m=+1202.481920192" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.301692 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2845-account-create-update-pzlvv" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.441756 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvp5z\" (UniqueName: \"kubernetes.io/projected/8d3720d6-e79c-44cc-8780-415f2b139a9f-kube-api-access-xvp5z\") pod \"8d3720d6-e79c-44cc-8780-415f2b139a9f\" (UID: \"8d3720d6-e79c-44cc-8780-415f2b139a9f\") " Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.442275 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d3720d6-e79c-44cc-8780-415f2b139a9f-operator-scripts\") pod \"8d3720d6-e79c-44cc-8780-415f2b139a9f\" (UID: \"8d3720d6-e79c-44cc-8780-415f2b139a9f\") " Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.442922 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d3720d6-e79c-44cc-8780-415f2b139a9f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8d3720d6-e79c-44cc-8780-415f2b139a9f" (UID: "8d3720d6-e79c-44cc-8780-415f2b139a9f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.449060 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d3720d6-e79c-44cc-8780-415f2b139a9f-kube-api-access-xvp5z" (OuterVolumeSpecName: "kube-api-access-xvp5z") pod "8d3720d6-e79c-44cc-8780-415f2b139a9f" (UID: "8d3720d6-e79c-44cc-8780-415f2b139a9f"). InnerVolumeSpecName "kube-api-access-xvp5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.501005 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6479-account-create-update-5jbnw" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.506376 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vphqv" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.511901 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lgxt2" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.519999 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tvmcg" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.530635 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2e37-account-create-update-dxdg7" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.543917 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvp5z\" (UniqueName: \"kubernetes.io/projected/8d3720d6-e79c-44cc-8780-415f2b139a9f-kube-api-access-xvp5z\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.543948 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d3720d6-e79c-44cc-8780-415f2b139a9f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.645553 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8s5g8\" (UniqueName: \"kubernetes.io/projected/bd1f40c7-a082-4356-848b-578cb58f1015-kube-api-access-8s5g8\") pod \"bd1f40c7-a082-4356-848b-578cb58f1015\" (UID: \"bd1f40c7-a082-4356-848b-578cb58f1015\") " Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.645845 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p422l\" (UniqueName: \"kubernetes.io/projected/0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9-kube-api-access-p422l\") pod \"0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9\" (UID: \"0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9\") " Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.645897 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9-operator-scripts\") pod \"0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9\" (UID: \"0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9\") " Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.645925 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f-operator-scripts\") pod \"0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f\" (UID: \"0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f\") " Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.645951 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pkj7\" (UniqueName: \"kubernetes.io/projected/0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f-kube-api-access-8pkj7\") pod \"0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f\" (UID: \"0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f\") " Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.645969 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c2ab590-4b44-4d31-9324-2abffd328742-operator-scripts\") pod \"9c2ab590-4b44-4d31-9324-2abffd328742\" (UID: \"9c2ab590-4b44-4d31-9324-2abffd328742\") " Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.645998 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxszf\" (UniqueName: \"kubernetes.io/projected/b885ee2d-5c14-40b1-83c5-068d2bb0ef96-kube-api-access-pxszf\") pod \"b885ee2d-5c14-40b1-83c5-068d2bb0ef96\" (UID: \"b885ee2d-5c14-40b1-83c5-068d2bb0ef96\") " Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.646019 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffwbn\" (UniqueName: \"kubernetes.io/projected/9c2ab590-4b44-4d31-9324-2abffd328742-kube-api-access-ffwbn\") pod \"9c2ab590-4b44-4d31-9324-2abffd328742\" (UID: \"9c2ab590-4b44-4d31-9324-2abffd328742\") " Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.646039 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd1f40c7-a082-4356-848b-578cb58f1015-operator-scripts\") pod \"bd1f40c7-a082-4356-848b-578cb58f1015\" (UID: \"bd1f40c7-a082-4356-848b-578cb58f1015\") " Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.646098 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b885ee2d-5c14-40b1-83c5-068d2bb0ef96-operator-scripts\") pod \"b885ee2d-5c14-40b1-83c5-068d2bb0ef96\" (UID: \"b885ee2d-5c14-40b1-83c5-068d2bb0ef96\") " Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.646457 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f" (UID: "0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.646654 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9" (UID: "0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.646791 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd1f40c7-a082-4356-848b-578cb58f1015-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bd1f40c7-a082-4356-848b-578cb58f1015" (UID: "bd1f40c7-a082-4356-848b-578cb58f1015"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.646874 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c2ab590-4b44-4d31-9324-2abffd328742-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9c2ab590-4b44-4d31-9324-2abffd328742" (UID: "9c2ab590-4b44-4d31-9324-2abffd328742"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.646895 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b885ee2d-5c14-40b1-83c5-068d2bb0ef96-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b885ee2d-5c14-40b1-83c5-068d2bb0ef96" (UID: "b885ee2d-5c14-40b1-83c5-068d2bb0ef96"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.647401 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.647427 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.647440 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c2ab590-4b44-4d31-9324-2abffd328742-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.647454 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd1f40c7-a082-4356-848b-578cb58f1015-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.647465 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b885ee2d-5c14-40b1-83c5-068d2bb0ef96-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.649716 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9-kube-api-access-p422l" (OuterVolumeSpecName: "kube-api-access-p422l") pod "0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9" (UID: "0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9"). InnerVolumeSpecName "kube-api-access-p422l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.650516 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c2ab590-4b44-4d31-9324-2abffd328742-kube-api-access-ffwbn" (OuterVolumeSpecName: "kube-api-access-ffwbn") pod "9c2ab590-4b44-4d31-9324-2abffd328742" (UID: "9c2ab590-4b44-4d31-9324-2abffd328742"). InnerVolumeSpecName "kube-api-access-ffwbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.650617 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd1f40c7-a082-4356-848b-578cb58f1015-kube-api-access-8s5g8" (OuterVolumeSpecName: "kube-api-access-8s5g8") pod "bd1f40c7-a082-4356-848b-578cb58f1015" (UID: "bd1f40c7-a082-4356-848b-578cb58f1015"). InnerVolumeSpecName "kube-api-access-8s5g8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.650666 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b885ee2d-5c14-40b1-83c5-068d2bb0ef96-kube-api-access-pxszf" (OuterVolumeSpecName: "kube-api-access-pxszf") pod "b885ee2d-5c14-40b1-83c5-068d2bb0ef96" (UID: "b885ee2d-5c14-40b1-83c5-068d2bb0ef96"). InnerVolumeSpecName "kube-api-access-pxszf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.650906 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f-kube-api-access-8pkj7" (OuterVolumeSpecName: "kube-api-access-8pkj7") pod "0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f" (UID: "0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f"). InnerVolumeSpecName "kube-api-access-8pkj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.748974 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8s5g8\" (UniqueName: \"kubernetes.io/projected/bd1f40c7-a082-4356-848b-578cb58f1015-kube-api-access-8s5g8\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.749004 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p422l\" (UniqueName: \"kubernetes.io/projected/0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9-kube-api-access-p422l\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.749014 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pkj7\" (UniqueName: \"kubernetes.io/projected/0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f-kube-api-access-8pkj7\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.749023 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxszf\" (UniqueName: \"kubernetes.io/projected/b885ee2d-5c14-40b1-83c5-068d2bb0ef96-kube-api-access-pxszf\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.749032 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffwbn\" (UniqueName: \"kubernetes.io/projected/9c2ab590-4b44-4d31-9324-2abffd328742-kube-api-access-ffwbn\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.948970 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tvmcg" event={"ID":"b885ee2d-5c14-40b1-83c5-068d2bb0ef96","Type":"ContainerDied","Data":"f08c620f9738d9d005efb3e0f8785bed538d2be949e2275b3b0d3dbc7f900b23"} Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.949024 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f08c620f9738d9d005efb3e0f8785bed538d2be949e2275b3b0d3dbc7f900b23" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.949032 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tvmcg" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.951927 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6479-account-create-update-5jbnw" event={"ID":"0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f","Type":"ContainerDied","Data":"dd2544d6c76e4508fc55b06d3ecd383b8b4524287ac4abf4ea2650248ae103bd"} Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.951966 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6479-account-create-update-5jbnw" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.951981 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd2544d6c76e4508fc55b06d3ecd383b8b4524287ac4abf4ea2650248ae103bd" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.953410 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lgxt2" event={"ID":"9c2ab590-4b44-4d31-9324-2abffd328742","Type":"ContainerDied","Data":"33f1a9fd25989044afe10558bdf23bdd2bc2ff0fea1df6026eb0d27274d133cf"} Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.953458 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33f1a9fd25989044afe10558bdf23bdd2bc2ff0fea1df6026eb0d27274d133cf" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.953527 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lgxt2" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.961367 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vphqv" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.961394 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vphqv" event={"ID":"0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9","Type":"ContainerDied","Data":"9912226f1b612f4d9b536d7b400ccde25f68c67215bb6e3c9397cbe63c6b3573"} Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.961447 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9912226f1b612f4d9b536d7b400ccde25f68c67215bb6e3c9397cbe63c6b3573" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.963180 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2845-account-create-update-pzlvv" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.963178 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2845-account-create-update-pzlvv" event={"ID":"8d3720d6-e79c-44cc-8780-415f2b139a9f","Type":"ContainerDied","Data":"16cde31a3a4e1294bd73e9bf2081447eeaca3c29926f59aea5ed8b06a508379e"} Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.963305 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16cde31a3a4e1294bd73e9bf2081447eeaca3c29926f59aea5ed8b06a508379e" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.964724 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2e37-account-create-update-dxdg7" event={"ID":"bd1f40c7-a082-4356-848b-578cb58f1015","Type":"ContainerDied","Data":"65804bb23dea7d31d2b25d27353173fdd535d2a5c13254fbe04b194ee5250fb9"} Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.964760 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65804bb23dea7d31d2b25d27353173fdd535d2a5c13254fbe04b194ee5250fb9" Feb 27 16:43:11 crc kubenswrapper[4814]: I0227 16:43:11.964822 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2e37-account-create-update-dxdg7" Feb 27 16:43:14 crc kubenswrapper[4814]: I0227 16:43:14.996640 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-l4sht" event={"ID":"6779899f-08d7-4de4-884c-070af4597386","Type":"ContainerStarted","Data":"76d4b032aba40b34faa39cbb6776ef5e48eedd9b1974b40a8516c0075c6b2d3c"} Feb 27 16:43:15 crc kubenswrapper[4814]: I0227 16:43:15.013104 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-l4sht" podStartSLOduration=1.9271325780000002 podStartE2EDuration="7.01306534s" podCreationTimestamp="2026-02-27 16:43:08 +0000 UTC" firstStartedPulling="2026-02-27 16:43:09.270524777 +0000 UTC m=+1201.723149607" lastFinishedPulling="2026-02-27 16:43:14.356457489 +0000 UTC m=+1206.809082369" observedRunningTime="2026-02-27 16:43:15.012251015 +0000 UTC m=+1207.464875865" watchObservedRunningTime="2026-02-27 16:43:15.01306534 +0000 UTC m=+1207.465690210" Feb 27 16:43:16 crc kubenswrapper[4814]: I0227 16:43:16.800866 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:16 crc kubenswrapper[4814]: I0227 16:43:16.880108 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-p6sq9"] Feb 27 16:43:16 crc kubenswrapper[4814]: I0227 16:43:16.880396 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-p6sq9" podUID="958486c1-15d1-4184-9197-84fd9c0ba31b" containerName="dnsmasq-dns" containerID="cri-o://6669b50d32f8a4e6280bb46209b22c07a78970e4deecb6329c65ddeeaea0a094" gracePeriod=10 Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.016657 4814 generic.go:334] "Generic (PLEG): container finished" podID="958486c1-15d1-4184-9197-84fd9c0ba31b" containerID="6669b50d32f8a4e6280bb46209b22c07a78970e4deecb6329c65ddeeaea0a094" exitCode=0 Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.016704 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-p6sq9" event={"ID":"958486c1-15d1-4184-9197-84fd9c0ba31b","Type":"ContainerDied","Data":"6669b50d32f8a4e6280bb46209b22c07a78970e4deecb6329c65ddeeaea0a094"} Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.102589 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-p6sq9" podUID="958486c1-15d1-4184-9197-84fd9c0ba31b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.376341 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.450146 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-dns-svc\") pod \"958486c1-15d1-4184-9197-84fd9c0ba31b\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.450203 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-ovsdbserver-sb\") pod \"958486c1-15d1-4184-9197-84fd9c0ba31b\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.450229 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-config\") pod \"958486c1-15d1-4184-9197-84fd9c0ba31b\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.450378 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-ovsdbserver-nb\") pod \"958486c1-15d1-4184-9197-84fd9c0ba31b\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.450398 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hx4vj\" (UniqueName: \"kubernetes.io/projected/958486c1-15d1-4184-9197-84fd9c0ba31b-kube-api-access-hx4vj\") pod \"958486c1-15d1-4184-9197-84fd9c0ba31b\" (UID: \"958486c1-15d1-4184-9197-84fd9c0ba31b\") " Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.457567 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/958486c1-15d1-4184-9197-84fd9c0ba31b-kube-api-access-hx4vj" (OuterVolumeSpecName: "kube-api-access-hx4vj") pod "958486c1-15d1-4184-9197-84fd9c0ba31b" (UID: "958486c1-15d1-4184-9197-84fd9c0ba31b"). InnerVolumeSpecName "kube-api-access-hx4vj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.493653 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "958486c1-15d1-4184-9197-84fd9c0ba31b" (UID: "958486c1-15d1-4184-9197-84fd9c0ba31b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.514757 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "958486c1-15d1-4184-9197-84fd9c0ba31b" (UID: "958486c1-15d1-4184-9197-84fd9c0ba31b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.515817 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "958486c1-15d1-4184-9197-84fd9c0ba31b" (UID: "958486c1-15d1-4184-9197-84fd9c0ba31b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.533117 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-config" (OuterVolumeSpecName: "config") pod "958486c1-15d1-4184-9197-84fd9c0ba31b" (UID: "958486c1-15d1-4184-9197-84fd9c0ba31b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.552636 4814 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.552672 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.552686 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.552697 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/958486c1-15d1-4184-9197-84fd9c0ba31b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:17 crc kubenswrapper[4814]: I0227 16:43:17.552709 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hx4vj\" (UniqueName: \"kubernetes.io/projected/958486c1-15d1-4184-9197-84fd9c0ba31b-kube-api-access-hx4vj\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:18 crc kubenswrapper[4814]: I0227 16:43:18.028064 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-p6sq9" event={"ID":"958486c1-15d1-4184-9197-84fd9c0ba31b","Type":"ContainerDied","Data":"abb6015521fefb10a106eea8fc30e0578835eca45614bb6b121b64214432cc2a"} Feb 27 16:43:18 crc kubenswrapper[4814]: I0227 16:43:18.028491 4814 scope.go:117] "RemoveContainer" containerID="6669b50d32f8a4e6280bb46209b22c07a78970e4deecb6329c65ddeeaea0a094" Feb 27 16:43:18 crc kubenswrapper[4814]: I0227 16:43:18.028093 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-p6sq9" Feb 27 16:43:18 crc kubenswrapper[4814]: I0227 16:43:18.031034 4814 generic.go:334] "Generic (PLEG): container finished" podID="6779899f-08d7-4de4-884c-070af4597386" containerID="76d4b032aba40b34faa39cbb6776ef5e48eedd9b1974b40a8516c0075c6b2d3c" exitCode=0 Feb 27 16:43:18 crc kubenswrapper[4814]: I0227 16:43:18.031073 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-l4sht" event={"ID":"6779899f-08d7-4de4-884c-070af4597386","Type":"ContainerDied","Data":"76d4b032aba40b34faa39cbb6776ef5e48eedd9b1974b40a8516c0075c6b2d3c"} Feb 27 16:43:18 crc kubenswrapper[4814]: I0227 16:43:18.052649 4814 scope.go:117] "RemoveContainer" containerID="749ff4565088119df234b31211740010acefc4e5ed9665a0e594cb14b4cf5724" Feb 27 16:43:18 crc kubenswrapper[4814]: I0227 16:43:18.086895 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-p6sq9"] Feb 27 16:43:18 crc kubenswrapper[4814]: I0227 16:43:18.096936 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-p6sq9"] Feb 27 16:43:18 crc kubenswrapper[4814]: I0227 16:43:18.499231 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="958486c1-15d1-4184-9197-84fd9c0ba31b" path="/var/lib/kubelet/pods/958486c1-15d1-4184-9197-84fd9c0ba31b/volumes" Feb 27 16:43:19 crc kubenswrapper[4814]: I0227 16:43:19.371723 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-l4sht" Feb 27 16:43:19 crc kubenswrapper[4814]: I0227 16:43:19.487491 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6779899f-08d7-4de4-884c-070af4597386-config-data\") pod \"6779899f-08d7-4de4-884c-070af4597386\" (UID: \"6779899f-08d7-4de4-884c-070af4597386\") " Feb 27 16:43:19 crc kubenswrapper[4814]: I0227 16:43:19.487602 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6779899f-08d7-4de4-884c-070af4597386-combined-ca-bundle\") pod \"6779899f-08d7-4de4-884c-070af4597386\" (UID: \"6779899f-08d7-4de4-884c-070af4597386\") " Feb 27 16:43:19 crc kubenswrapper[4814]: I0227 16:43:19.487653 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4q2c\" (UniqueName: \"kubernetes.io/projected/6779899f-08d7-4de4-884c-070af4597386-kube-api-access-t4q2c\") pod \"6779899f-08d7-4de4-884c-070af4597386\" (UID: \"6779899f-08d7-4de4-884c-070af4597386\") " Feb 27 16:43:19 crc kubenswrapper[4814]: I0227 16:43:19.499575 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6779899f-08d7-4de4-884c-070af4597386-kube-api-access-t4q2c" (OuterVolumeSpecName: "kube-api-access-t4q2c") pod "6779899f-08d7-4de4-884c-070af4597386" (UID: "6779899f-08d7-4de4-884c-070af4597386"). InnerVolumeSpecName "kube-api-access-t4q2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:19 crc kubenswrapper[4814]: I0227 16:43:19.519412 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6779899f-08d7-4de4-884c-070af4597386-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6779899f-08d7-4de4-884c-070af4597386" (UID: "6779899f-08d7-4de4-884c-070af4597386"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:19 crc kubenswrapper[4814]: I0227 16:43:19.563583 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6779899f-08d7-4de4-884c-070af4597386-config-data" (OuterVolumeSpecName: "config-data") pod "6779899f-08d7-4de4-884c-070af4597386" (UID: "6779899f-08d7-4de4-884c-070af4597386"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:19 crc kubenswrapper[4814]: I0227 16:43:19.590409 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6779899f-08d7-4de4-884c-070af4597386-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:19 crc kubenswrapper[4814]: I0227 16:43:19.590465 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4q2c\" (UniqueName: \"kubernetes.io/projected/6779899f-08d7-4de4-884c-070af4597386-kube-api-access-t4q2c\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:19 crc kubenswrapper[4814]: I0227 16:43:19.590486 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6779899f-08d7-4de4-884c-070af4597386-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.053077 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-l4sht" event={"ID":"6779899f-08d7-4de4-884c-070af4597386","Type":"ContainerDied","Data":"6a5592faccbb78c9b6af076be3533a9c1ea9fd5baef513911445df24b8b0addf"} Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.053131 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a5592faccbb78c9b6af076be3533a9c1ea9fd5baef513911445df24b8b0addf" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.053155 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-l4sht" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.347950 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs"] Feb 27 16:43:20 crc kubenswrapper[4814]: E0227 16:43:20.348712 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd1f40c7-a082-4356-848b-578cb58f1015" containerName="mariadb-account-create-update" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.348741 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd1f40c7-a082-4356-848b-578cb58f1015" containerName="mariadb-account-create-update" Feb 27 16:43:20 crc kubenswrapper[4814]: E0227 16:43:20.348754 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d3720d6-e79c-44cc-8780-415f2b139a9f" containerName="mariadb-account-create-update" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.348764 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d3720d6-e79c-44cc-8780-415f2b139a9f" containerName="mariadb-account-create-update" Feb 27 16:43:20 crc kubenswrapper[4814]: E0227 16:43:20.348792 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="958486c1-15d1-4184-9197-84fd9c0ba31b" containerName="dnsmasq-dns" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.348802 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="958486c1-15d1-4184-9197-84fd9c0ba31b" containerName="dnsmasq-dns" Feb 27 16:43:20 crc kubenswrapper[4814]: E0227 16:43:20.348820 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f" containerName="mariadb-account-create-update" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.348828 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f" containerName="mariadb-account-create-update" Feb 27 16:43:20 crc kubenswrapper[4814]: E0227 16:43:20.348853 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6779899f-08d7-4de4-884c-070af4597386" containerName="keystone-db-sync" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.348861 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="6779899f-08d7-4de4-884c-070af4597386" containerName="keystone-db-sync" Feb 27 16:43:20 crc kubenswrapper[4814]: E0227 16:43:20.348874 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c2ab590-4b44-4d31-9324-2abffd328742" containerName="mariadb-database-create" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.348882 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c2ab590-4b44-4d31-9324-2abffd328742" containerName="mariadb-database-create" Feb 27 16:43:20 crc kubenswrapper[4814]: E0227 16:43:20.348901 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="958486c1-15d1-4184-9197-84fd9c0ba31b" containerName="init" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.348909 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="958486c1-15d1-4184-9197-84fd9c0ba31b" containerName="init" Feb 27 16:43:20 crc kubenswrapper[4814]: E0227 16:43:20.348924 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b885ee2d-5c14-40b1-83c5-068d2bb0ef96" containerName="mariadb-database-create" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.348932 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="b885ee2d-5c14-40b1-83c5-068d2bb0ef96" containerName="mariadb-database-create" Feb 27 16:43:20 crc kubenswrapper[4814]: E0227 16:43:20.348945 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9" containerName="mariadb-database-create" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.348952 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9" containerName="mariadb-database-create" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.349642 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9" containerName="mariadb-database-create" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.349686 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c2ab590-4b44-4d31-9324-2abffd328742" containerName="mariadb-database-create" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.349701 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="958486c1-15d1-4184-9197-84fd9c0ba31b" containerName="dnsmasq-dns" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.349712 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d3720d6-e79c-44cc-8780-415f2b139a9f" containerName="mariadb-account-create-update" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.349722 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="b885ee2d-5c14-40b1-83c5-068d2bb0ef96" containerName="mariadb-database-create" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.349734 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd1f40c7-a082-4356-848b-578cb58f1015" containerName="mariadb-account-create-update" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.349746 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f" containerName="mariadb-account-create-update" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.349758 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="6779899f-08d7-4de4-884c-070af4597386" containerName="keystone-db-sync" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.350804 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.372396 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-bxvjh"] Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.373597 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.380368 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.380790 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-m5bsz" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.381053 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.381374 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.381626 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs"] Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.383057 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.407837 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.407909 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97qz8\" (UniqueName: \"kubernetes.io/projected/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-kube-api-access-97qz8\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.407964 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.407991 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.408039 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-config\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.408076 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.415007 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bxvjh"] Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.512378 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.512451 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-combined-ca-bundle\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.512476 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.512512 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97qz8\" (UniqueName: \"kubernetes.io/projected/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-kube-api-access-97qz8\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.512539 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-scripts\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.512557 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-credential-keys\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.512581 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.512601 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.512621 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-config-data\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.512643 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jm9x\" (UniqueName: \"kubernetes.io/projected/1c68eb62-1f62-4ed1-a2cb-405cda573e09-kube-api-access-7jm9x\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.512668 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-config\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.512684 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-fernet-keys\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.513461 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.513959 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.514688 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.515171 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.515696 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-config\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.533167 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5c887c6797-kst67"] Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.534363 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.541730 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.542115 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-drqm2" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.542199 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.543786 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.554276 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5c887c6797-kst67"] Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.569204 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97qz8\" (UniqueName: \"kubernetes.io/projected/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-kube-api-access-97qz8\") pod \"dnsmasq-dns-5c5cc7c5ff-7dhfs\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.596499 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-vqkl5"] Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.597750 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.612555 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.612797 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-89btm" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.612707 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.614292 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-credential-keys\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.614418 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9lgb\" (UniqueName: \"kubernetes.io/projected/3a57db6d-9660-4c37-847d-b25c6d65e147-kube-api-access-k9lgb\") pod \"horizon-5c887c6797-kst67\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.614487 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3a57db6d-9660-4c37-847d-b25c6d65e147-horizon-secret-key\") pod \"horizon-5c887c6797-kst67\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.614555 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-config-data\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.614615 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a57db6d-9660-4c37-847d-b25c6d65e147-config-data\") pod \"horizon-5c887c6797-kst67\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.614689 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jm9x\" (UniqueName: \"kubernetes.io/projected/1c68eb62-1f62-4ed1-a2cb-405cda573e09-kube-api-access-7jm9x\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.614767 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-fernet-keys\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.614907 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-combined-ca-bundle\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.615030 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a57db6d-9660-4c37-847d-b25c6d65e147-scripts\") pod \"horizon-5c887c6797-kst67\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.615103 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-scripts\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.615167 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a57db6d-9660-4c37-847d-b25c6d65e147-logs\") pod \"horizon-5c887c6797-kst67\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.617176 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-credential-keys\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.624968 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-combined-ca-bundle\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.625575 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-config-data\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.639157 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-fernet-keys\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.641581 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-scripts\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.650455 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vqkl5"] Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.651063 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jm9x\" (UniqueName: \"kubernetes.io/projected/1c68eb62-1f62-4ed1-a2cb-405cda573e09-kube-api-access-7jm9x\") pod \"keystone-bootstrap-bxvjh\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.667330 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.669324 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.683023 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.683812 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.695031 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.695353 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.716920 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a57db6d-9660-4c37-847d-b25c6d65e147-scripts\") pod \"horizon-5c887c6797-kst67\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.716978 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a57db6d-9660-4c37-847d-b25c6d65e147-logs\") pod \"horizon-5c887c6797-kst67\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.717021 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9lgb\" (UniqueName: \"kubernetes.io/projected/3a57db6d-9660-4c37-847d-b25c6d65e147-kube-api-access-k9lgb\") pod \"horizon-5c887c6797-kst67\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.717042 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3a57db6d-9660-4c37-847d-b25c6d65e147-horizon-secret-key\") pod \"horizon-5c887c6797-kst67\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.717062 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a57db6d-9660-4c37-847d-b25c6d65e147-config-data\") pod \"horizon-5c887c6797-kst67\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.718263 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a57db6d-9660-4c37-847d-b25c6d65e147-logs\") pod \"horizon-5c887c6797-kst67\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.718348 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a57db6d-9660-4c37-847d-b25c6d65e147-config-data\") pod \"horizon-5c887c6797-kst67\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.718797 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a57db6d-9660-4c37-847d-b25c6d65e147-scripts\") pod \"horizon-5c887c6797-kst67\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.728371 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-th62b"] Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.729327 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-th62b" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.748155 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.749559 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.750367 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-lhq6v" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.750421 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.750912 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3a57db6d-9660-4c37-847d-b25c6d65e147-horizon-secret-key\") pod \"horizon-5c887c6797-kst67\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.771106 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-th62b"] Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.773112 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9lgb\" (UniqueName: \"kubernetes.io/projected/3a57db6d-9660-4c37-847d-b25c6d65e147-kube-api-access-k9lgb\") pod \"horizon-5c887c6797-kst67\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.818017 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-run-httpd\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.818066 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-etc-machine-id\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.818098 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.818118 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-log-httpd\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.818182 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-combined-ca-bundle\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.818199 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-db-sync-config-data\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.818216 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdqh8\" (UniqueName: \"kubernetes.io/projected/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-kube-api-access-fdqh8\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.818233 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-scripts\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.818249 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-scripts\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.818285 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpmlb\" (UniqueName: \"kubernetes.io/projected/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-kube-api-access-rpmlb\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.818309 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-config-data\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.818332 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-config-data\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.818431 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.842984 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6ff5f79df7-n8ks6"] Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.844303 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.852622 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.862243 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6ff5f79df7-n8ks6"] Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.904954 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-pm95c"] Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.906100 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pm95c" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.913505 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-ll2qr" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.917244 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.920093 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6xn5\" (UniqueName: \"kubernetes.io/projected/5e413768-6d50-4725-a809-3658e913cfed-kube-api-access-m6xn5\") pod \"neutron-db-sync-th62b\" (UID: \"5e413768-6d50-4725-a809-3658e913cfed\") " pod="openstack/neutron-db-sync-th62b" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.920152 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e413768-6d50-4725-a809-3658e913cfed-combined-ca-bundle\") pod \"neutron-db-sync-th62b\" (UID: \"5e413768-6d50-4725-a809-3658e913cfed\") " pod="openstack/neutron-db-sync-th62b" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.920179 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-combined-ca-bundle\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.920198 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-db-sync-config-data\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.920217 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdqh8\" (UniqueName: \"kubernetes.io/projected/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-kube-api-access-fdqh8\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.920238 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-scripts\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.920271 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-scripts\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.920295 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpmlb\" (UniqueName: \"kubernetes.io/projected/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-kube-api-access-rpmlb\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.920316 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-config-data\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.920338 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-config-data\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.920365 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.920389 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-run-httpd\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.920410 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-etc-machine-id\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.920435 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.920452 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-log-httpd\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.920477 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5e413768-6d50-4725-a809-3658e913cfed-config\") pod \"neutron-db-sync-th62b\" (UID: \"5e413768-6d50-4725-a809-3658e913cfed\") " pod="openstack/neutron-db-sync-th62b" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.925932 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-etc-machine-id\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.952360 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-run-httpd\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.955898 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-scripts\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.956302 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-db-sync-config-data\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.957101 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-config-data\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.957633 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-combined-ca-bundle\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.961443 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-pq8r4"] Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.962029 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-log-httpd\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.962566 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-scripts\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.963245 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.967287 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.969479 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-config-data\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.972127 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.979922 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.980275 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-k25g2" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.982724 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.982933 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pm95c"] Feb 27 16:43:20 crc kubenswrapper[4814]: I0227 16:43:20.994240 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-pq8r4"] Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.015150 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdqh8\" (UniqueName: \"kubernetes.io/projected/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-kube-api-access-fdqh8\") pod \"ceilometer-0\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " pod="openstack/ceilometer-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.026387 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs"] Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.029770 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/548a8aff-dc75-4e9f-857a-0acab993607c-logs\") pod \"horizon-6ff5f79df7-n8ks6\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.030557 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-config-data\") pod \"placement-db-sync-pq8r4\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.030662 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6xn5\" (UniqueName: \"kubernetes.io/projected/5e413768-6d50-4725-a809-3658e913cfed-kube-api-access-m6xn5\") pod \"neutron-db-sync-th62b\" (UID: \"5e413768-6d50-4725-a809-3658e913cfed\") " pod="openstack/neutron-db-sync-th62b" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.030739 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-combined-ca-bundle\") pod \"placement-db-sync-pq8r4\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.030862 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e413768-6d50-4725-a809-3658e913cfed-combined-ca-bundle\") pod \"neutron-db-sync-th62b\" (UID: \"5e413768-6d50-4725-a809-3658e913cfed\") " pod="openstack/neutron-db-sync-th62b" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.030964 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-db-sync-config-data\") pod \"barbican-db-sync-pm95c\" (UID: \"c357e0a2-9571-4e9e-b48c-0f92e2c99afc\") " pod="openstack/barbican-db-sync-pm95c" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.036595 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk6b4\" (UniqueName: \"kubernetes.io/projected/548a8aff-dc75-4e9f-857a-0acab993607c-kube-api-access-mk6b4\") pod \"horizon-6ff5f79df7-n8ks6\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.036690 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/548a8aff-dc75-4e9f-857a-0acab993607c-config-data\") pod \"horizon-6ff5f79df7-n8ks6\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.036781 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-scripts\") pod \"placement-db-sync-pq8r4\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.036820 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/548a8aff-dc75-4e9f-857a-0acab993607c-scripts\") pod \"horizon-6ff5f79df7-n8ks6\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.036867 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/548a8aff-dc75-4e9f-857a-0acab993607c-horizon-secret-key\") pod \"horizon-6ff5f79df7-n8ks6\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.036942 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-combined-ca-bundle\") pod \"barbican-db-sync-pm95c\" (UID: \"c357e0a2-9571-4e9e-b48c-0f92e2c99afc\") " pod="openstack/barbican-db-sync-pm95c" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.037015 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f5812c2-9952-4125-b055-dd2888fad3f5-logs\") pod \"placement-db-sync-pq8r4\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.037019 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpmlb\" (UniqueName: \"kubernetes.io/projected/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-kube-api-access-rpmlb\") pod \"cinder-db-sync-vqkl5\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.037141 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsd8h\" (UniqueName: \"kubernetes.io/projected/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-kube-api-access-qsd8h\") pod \"barbican-db-sync-pm95c\" (UID: \"c357e0a2-9571-4e9e-b48c-0f92e2c99afc\") " pod="openstack/barbican-db-sync-pm95c" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.037171 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5e413768-6d50-4725-a809-3658e913cfed-config\") pod \"neutron-db-sync-th62b\" (UID: \"5e413768-6d50-4725-a809-3658e913cfed\") " pod="openstack/neutron-db-sync-th62b" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.037298 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k72fh\" (UniqueName: \"kubernetes.io/projected/6f5812c2-9952-4125-b055-dd2888fad3f5-kube-api-access-k72fh\") pod \"placement-db-sync-pq8r4\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.037558 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e413768-6d50-4725-a809-3658e913cfed-combined-ca-bundle\") pod \"neutron-db-sync-th62b\" (UID: \"5e413768-6d50-4725-a809-3658e913cfed\") " pod="openstack/neutron-db-sync-th62b" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.041911 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5e413768-6d50-4725-a809-3658e913cfed-config\") pod \"neutron-db-sync-th62b\" (UID: \"5e413768-6d50-4725-a809-3658e913cfed\") " pod="openstack/neutron-db-sync-th62b" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.059073 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.070360 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-rv8p6"] Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.071900 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6xn5\" (UniqueName: \"kubernetes.io/projected/5e413768-6d50-4725-a809-3658e913cfed-kube-api-access-m6xn5\") pod \"neutron-db-sync-th62b\" (UID: \"5e413768-6d50-4725-a809-3658e913cfed\") " pod="openstack/neutron-db-sync-th62b" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.072012 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.130519 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138530 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-combined-ca-bundle\") pod \"barbican-db-sync-pm95c\" (UID: \"c357e0a2-9571-4e9e-b48c-0f92e2c99afc\") " pod="openstack/barbican-db-sync-pm95c" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138589 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f5812c2-9952-4125-b055-dd2888fad3f5-logs\") pod \"placement-db-sync-pq8r4\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138617 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd7n5\" (UniqueName: \"kubernetes.io/projected/93476b12-7c88-4276-b58f-2a6173a7eca1-kube-api-access-vd7n5\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138656 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsd8h\" (UniqueName: \"kubernetes.io/projected/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-kube-api-access-qsd8h\") pod \"barbican-db-sync-pm95c\" (UID: \"c357e0a2-9571-4e9e-b48c-0f92e2c99afc\") " pod="openstack/barbican-db-sync-pm95c" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138686 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k72fh\" (UniqueName: \"kubernetes.io/projected/6f5812c2-9952-4125-b055-dd2888fad3f5-kube-api-access-k72fh\") pod \"placement-db-sync-pq8r4\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138708 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/548a8aff-dc75-4e9f-857a-0acab993607c-logs\") pod \"horizon-6ff5f79df7-n8ks6\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138723 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-config-data\") pod \"placement-db-sync-pq8r4\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138747 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-combined-ca-bundle\") pod \"placement-db-sync-pq8r4\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138763 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-config\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138796 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-db-sync-config-data\") pod \"barbican-db-sync-pm95c\" (UID: \"c357e0a2-9571-4e9e-b48c-0f92e2c99afc\") " pod="openstack/barbican-db-sync-pm95c" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138813 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138830 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk6b4\" (UniqueName: \"kubernetes.io/projected/548a8aff-dc75-4e9f-857a-0acab993607c-kube-api-access-mk6b4\") pod \"horizon-6ff5f79df7-n8ks6\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138852 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/548a8aff-dc75-4e9f-857a-0acab993607c-config-data\") pod \"horizon-6ff5f79df7-n8ks6\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138871 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138888 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138904 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-scripts\") pod \"placement-db-sync-pq8r4\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138924 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/548a8aff-dc75-4e9f-857a-0acab993607c-scripts\") pod \"horizon-6ff5f79df7-n8ks6\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138942 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.138961 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/548a8aff-dc75-4e9f-857a-0acab993607c-horizon-secret-key\") pod \"horizon-6ff5f79df7-n8ks6\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.139338 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f5812c2-9952-4125-b055-dd2888fad3f5-logs\") pod \"placement-db-sync-pq8r4\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.139706 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/548a8aff-dc75-4e9f-857a-0acab993607c-logs\") pod \"horizon-6ff5f79df7-n8ks6\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.141208 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/548a8aff-dc75-4e9f-857a-0acab993607c-config-data\") pod \"horizon-6ff5f79df7-n8ks6\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.143525 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/548a8aff-dc75-4e9f-857a-0acab993607c-scripts\") pod \"horizon-6ff5f79df7-n8ks6\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.154650 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-th62b" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.160457 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-db-sync-config-data\") pod \"barbican-db-sync-pm95c\" (UID: \"c357e0a2-9571-4e9e-b48c-0f92e2c99afc\") " pod="openstack/barbican-db-sync-pm95c" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.160747 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-scripts\") pod \"placement-db-sync-pq8r4\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.161313 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-combined-ca-bundle\") pod \"placement-db-sync-pq8r4\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.168816 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.180427 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.181315 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-config-data\") pod \"placement-db-sync-pq8r4\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.181739 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-rv8p6"] Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.186222 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk6b4\" (UniqueName: \"kubernetes.io/projected/548a8aff-dc75-4e9f-857a-0acab993607c-kube-api-access-mk6b4\") pod \"horizon-6ff5f79df7-n8ks6\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.195859 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/548a8aff-dc75-4e9f-857a-0acab993607c-horizon-secret-key\") pod \"horizon-6ff5f79df7-n8ks6\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.200338 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.210844 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.215599 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsd8h\" (UniqueName: \"kubernetes.io/projected/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-kube-api-access-qsd8h\") pod \"barbican-db-sync-pm95c\" (UID: \"c357e0a2-9571-4e9e-b48c-0f92e2c99afc\") " pod="openstack/barbican-db-sync-pm95c" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.212953 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-dgrwq" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.213094 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.213149 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.221119 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k72fh\" (UniqueName: \"kubernetes.io/projected/6f5812c2-9952-4125-b055-dd2888fad3f5-kube-api-access-k72fh\") pod \"placement-db-sync-pq8r4\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.229848 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-combined-ca-bundle\") pod \"barbican-db-sync-pm95c\" (UID: \"c357e0a2-9571-4e9e-b48c-0f92e2c99afc\") " pod="openstack/barbican-db-sync-pm95c" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.253285 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.253389 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.253471 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.253741 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.253808 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-logs\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.253845 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd7n5\" (UniqueName: \"kubernetes.io/projected/93476b12-7c88-4276-b58f-2a6173a7eca1-kube-api-access-vd7n5\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.253915 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.253970 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9cj4\" (UniqueName: \"kubernetes.io/projected/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-kube-api-access-k9cj4\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.253992 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-scripts\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.254033 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-config-data\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.254801 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-config\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.254891 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.255004 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.255108 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.257124 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.257755 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.258283 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.265784 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.266112 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pm95c" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.268161 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-config\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.273940 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.278614 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.293072 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.293267 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.298774 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pq8r4" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.318305 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.319475 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd7n5\" (UniqueName: \"kubernetes.io/projected/93476b12-7c88-4276-b58f-2a6173a7eca1-kube-api-access-vd7n5\") pod \"dnsmasq-dns-8b5c85b87-rv8p6\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.358022 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-logs\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.358095 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.358128 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9cj4\" (UniqueName: \"kubernetes.io/projected/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-kube-api-access-k9cj4\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.358147 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-scripts\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.358165 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-config-data\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.358187 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.358285 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.358310 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.358745 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.359031 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.360237 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-logs\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.363246 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-scripts\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.363770 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.366894 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.386574 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-config-data\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.389436 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9cj4\" (UniqueName: \"kubernetes.io/projected/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-kube-api-access-k9cj4\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.408712 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.432967 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.461156 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.461208 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.461281 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3c3924ef-1488-4784-851a-5c7c2480fa00-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.461302 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.461330 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.461356 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.461382 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c3924ef-1488-4784-851a-5c7c2480fa00-logs\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.461421 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bnjq\" (UniqueName: \"kubernetes.io/projected/3c3924ef-1488-4784-851a-5c7c2480fa00-kube-api-access-6bnjq\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.484786 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.522988 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.567262 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bnjq\" (UniqueName: \"kubernetes.io/projected/3c3924ef-1488-4784-851a-5c7c2480fa00-kube-api-access-6bnjq\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.567702 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.567859 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.568177 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3c3924ef-1488-4784-851a-5c7c2480fa00-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.568357 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.569607 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.569868 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.570026 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c3924ef-1488-4784-851a-5c7c2480fa00-logs\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.570040 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.571755 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c3924ef-1488-4784-851a-5c7c2480fa00-logs\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.569269 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3c3924ef-1488-4784-851a-5c7c2480fa00-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.572860 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.573973 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.577423 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.612177 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.632509 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.642975 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bnjq\" (UniqueName: \"kubernetes.io/projected/3c3924ef-1488-4784-851a-5c7c2480fa00-kube-api-access-6bnjq\") pod \"glance-default-internal-api-0\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.660292 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.677672 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bxvjh"] Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.782803 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.798890 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs"] Feb 27 16:43:21 crc kubenswrapper[4814]: I0227 16:43:21.811004 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5c887c6797-kst67"] Feb 27 16:43:21 crc kubenswrapper[4814]: W0227 16:43:21.815219 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bb67c60_793f_4f2c_b8e5_11fb94a857b0.slice/crio-a765878dd3c51cbabd077aea39149e41e2c84945591c467aafbe5fbd424bb589 WatchSource:0}: Error finding container a765878dd3c51cbabd077aea39149e41e2c84945591c467aafbe5fbd424bb589: Status 404 returned error can't find the container with id a765878dd3c51cbabd077aea39149e41e2c84945591c467aafbe5fbd424bb589 Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.068877 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-th62b"] Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.082089 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pm95c"] Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.097922 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-th62b" event={"ID":"5e413768-6d50-4725-a809-3658e913cfed","Type":"ContainerStarted","Data":"426d5abb10ada29b603a0a679e0c83ebbd8aa1defd09e7c392d89480b81a779e"} Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.103886 4814 generic.go:334] "Generic (PLEG): container finished" podID="4bb67c60-793f-4f2c-b8e5-11fb94a857b0" containerID="4391d0e95e238ace1143e70ed0f5e2696cc7164c4cb2cd2482fe97e5679a27c3" exitCode=0 Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.103939 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" event={"ID":"4bb67c60-793f-4f2c-b8e5-11fb94a857b0","Type":"ContainerDied","Data":"4391d0e95e238ace1143e70ed0f5e2696cc7164c4cb2cd2482fe97e5679a27c3"} Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.103959 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" event={"ID":"4bb67c60-793f-4f2c-b8e5-11fb94a857b0","Type":"ContainerStarted","Data":"a765878dd3c51cbabd077aea39149e41e2c84945591c467aafbe5fbd424bb589"} Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.111851 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vqkl5"] Feb 27 16:43:22 crc kubenswrapper[4814]: W0227 16:43:22.120599 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc357e0a2_9571_4e9e_b48c_0f92e2c99afc.slice/crio-47c8485ae374b4e18fc8661595348d4ace5b324d6a797c81472eaf5e774aedc7 WatchSource:0}: Error finding container 47c8485ae374b4e18fc8661595348d4ace5b324d6a797c81472eaf5e774aedc7: Status 404 returned error can't find the container with id 47c8485ae374b4e18fc8661595348d4ace5b324d6a797c81472eaf5e774aedc7 Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.123311 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01245f0c-38fe-4f9a-93e5-c5c5d5f32259","Type":"ContainerStarted","Data":"f302b88507b3e57c745e98835c09044d94519da0717d49898477cae134d2a018"} Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.126026 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-pq8r4"] Feb 27 16:43:22 crc kubenswrapper[4814]: W0227 16:43:22.128929 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bcf9bc6_1fb4_47b4_b12d_017902e4992e.slice/crio-6ae9247a5eb96b5778623b9ad7c9cdb98e1e976b78d4caa4487eba75ce461606 WatchSource:0}: Error finding container 6ae9247a5eb96b5778623b9ad7c9cdb98e1e976b78d4caa4487eba75ce461606: Status 404 returned error can't find the container with id 6ae9247a5eb96b5778623b9ad7c9cdb98e1e976b78d4caa4487eba75ce461606 Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.134889 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bxvjh" event={"ID":"1c68eb62-1f62-4ed1-a2cb-405cda573e09","Type":"ContainerStarted","Data":"ba8e4d12b3a356f51ad9bcd93b46a3983571ebb22168d59fa156bd0aa103a861"} Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.134937 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bxvjh" event={"ID":"1c68eb62-1f62-4ed1-a2cb-405cda573e09","Type":"ContainerStarted","Data":"4234d0f926810b9d3c48e7263c539f370a012a2c0d502aa51105aa1af288c8dd"} Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.139933 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c887c6797-kst67" event={"ID":"3a57db6d-9660-4c37-847d-b25c6d65e147","Type":"ContainerStarted","Data":"0a03f5372e573e1f531baf20c83f62cb403192409cb87766fb4412843edf71c3"} Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.203172 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-bxvjh" podStartSLOduration=2.203151208 podStartE2EDuration="2.203151208s" podCreationTimestamp="2026-02-27 16:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:43:22.16049048 +0000 UTC m=+1214.613115310" watchObservedRunningTime="2026-02-27 16:43:22.203151208 +0000 UTC m=+1214.655776038" Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.303973 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6ff5f79df7-n8ks6"] Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.318483 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-rv8p6"] Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.519609 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.543754 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.709368 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-dns-svc\") pod \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.709451 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-config\") pod \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.709489 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97qz8\" (UniqueName: \"kubernetes.io/projected/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-kube-api-access-97qz8\") pod \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.709570 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-dns-swift-storage-0\") pod \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.709603 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-ovsdbserver-nb\") pod \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.709671 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-ovsdbserver-sb\") pod \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\" (UID: \"4bb67c60-793f-4f2c-b8e5-11fb94a857b0\") " Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.715102 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-kube-api-access-97qz8" (OuterVolumeSpecName: "kube-api-access-97qz8") pod "4bb67c60-793f-4f2c-b8e5-11fb94a857b0" (UID: "4bb67c60-793f-4f2c-b8e5-11fb94a857b0"). InnerVolumeSpecName "kube-api-access-97qz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.732985 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4bb67c60-793f-4f2c-b8e5-11fb94a857b0" (UID: "4bb67c60-793f-4f2c-b8e5-11fb94a857b0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.743011 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4bb67c60-793f-4f2c-b8e5-11fb94a857b0" (UID: "4bb67c60-793f-4f2c-b8e5-11fb94a857b0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.763934 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4bb67c60-793f-4f2c-b8e5-11fb94a857b0" (UID: "4bb67c60-793f-4f2c-b8e5-11fb94a857b0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.768411 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4bb67c60-793f-4f2c-b8e5-11fb94a857b0" (UID: "4bb67c60-793f-4f2c-b8e5-11fb94a857b0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.772790 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-config" (OuterVolumeSpecName: "config") pod "4bb67c60-793f-4f2c-b8e5-11fb94a857b0" (UID: "4bb67c60-793f-4f2c-b8e5-11fb94a857b0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.811968 4814 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.812003 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.812019 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97qz8\" (UniqueName: \"kubernetes.io/projected/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-kube-api-access-97qz8\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.812033 4814 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.812045 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:22 crc kubenswrapper[4814]: I0227 16:43:22.812055 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4bb67c60-793f-4f2c-b8e5-11fb94a857b0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.164201 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.196804 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pm95c" event={"ID":"c357e0a2-9571-4e9e-b48c-0f92e2c99afc","Type":"ContainerStarted","Data":"47c8485ae374b4e18fc8661595348d4ace5b324d6a797c81472eaf5e774aedc7"} Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.203114 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5c887c6797-kst67"] Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.207096 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"80158a87-46ee-48b4-8d92-6d97c7d3bcdb","Type":"ContainerStarted","Data":"d9711b917b1cf3291855dcb95258f508091299b101788ad3ce6a07e9e97bef97"} Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.220196 4814 generic.go:334] "Generic (PLEG): container finished" podID="93476b12-7c88-4276-b58f-2a6173a7eca1" containerID="ade4976a96e606bf2f3a5d54c59a659087beb45bf3c28e6f44af862ce1245604" exitCode=0 Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.220295 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" event={"ID":"93476b12-7c88-4276-b58f-2a6173a7eca1","Type":"ContainerDied","Data":"ade4976a96e606bf2f3a5d54c59a659087beb45bf3c28e6f44af862ce1245604"} Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.220323 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" event={"ID":"93476b12-7c88-4276-b58f-2a6173a7eca1","Type":"ContainerStarted","Data":"e130fb1fd885ffe92672fe9e9b0ee79f8573349034d6bb3a9ec9216a83215fb0"} Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.227448 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.263764 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-675cb4b9c-cgtnr"] Feb 27 16:43:23 crc kubenswrapper[4814]: E0227 16:43:23.269834 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bb67c60-793f-4f2c-b8e5-11fb94a857b0" containerName="init" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.270042 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bb67c60-793f-4f2c-b8e5-11fb94a857b0" containerName="init" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.270374 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bb67c60-793f-4f2c-b8e5-11fb94a857b0" containerName="init" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.273875 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.282792 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-675cb4b9c-cgtnr"] Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.295056 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-th62b" event={"ID":"5e413768-6d50-4725-a809-3658e913cfed","Type":"ContainerStarted","Data":"0615c5bc9eee0c9e2440a32cd8f263d45418c571909c1a6d783625924de3e53c"} Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.322440 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" event={"ID":"4bb67c60-793f-4f2c-b8e5-11fb94a857b0","Type":"ContainerDied","Data":"a765878dd3c51cbabd077aea39149e41e2c84945591c467aafbe5fbd424bb589"} Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.322496 4814 scope.go:117] "RemoveContainer" containerID="4391d0e95e238ace1143e70ed0f5e2696cc7164c4cb2cd2482fe97e5679a27c3" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.322621 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.360438 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vqkl5" event={"ID":"7bcf9bc6-1fb4-47b4-b12d-017902e4992e","Type":"ContainerStarted","Data":"6ae9247a5eb96b5778623b9ad7c9cdb98e1e976b78d4caa4487eba75ce461606"} Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.383843 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pq8r4" event={"ID":"6f5812c2-9952-4125-b055-dd2888fad3f5","Type":"ContainerStarted","Data":"f2680ac0ae84f0980dbfb19842ec22005549507baa3436332fa16fa6776a06b2"} Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.429380 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.435411 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-th62b" podStartSLOduration=3.435392581 podStartE2EDuration="3.435392581s" podCreationTimestamp="2026-02-27 16:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:43:23.386975462 +0000 UTC m=+1215.839600292" watchObservedRunningTime="2026-02-27 16:43:23.435392581 +0000 UTC m=+1215.888017411" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.460187 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73f2abe5-41a2-440e-8f09-be50549a403f-logs\") pod \"horizon-675cb4b9c-cgtnr\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.460240 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73f2abe5-41a2-440e-8f09-be50549a403f-config-data\") pod \"horizon-675cb4b9c-cgtnr\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.460281 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73f2abe5-41a2-440e-8f09-be50549a403f-scripts\") pod \"horizon-675cb4b9c-cgtnr\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.460302 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55cq5\" (UniqueName: \"kubernetes.io/projected/73f2abe5-41a2-440e-8f09-be50549a403f-kube-api-access-55cq5\") pod \"horizon-675cb4b9c-cgtnr\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.460368 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73f2abe5-41a2-440e-8f09-be50549a403f-horizon-secret-key\") pod \"horizon-675cb4b9c-cgtnr\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.467097 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs"] Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.474103 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ff5f79df7-n8ks6" event={"ID":"548a8aff-dc75-4e9f-857a-0acab993607c","Type":"ContainerStarted","Data":"e994f8368bf85075374a41b11ee08168333adf6ee2e00a348c3a983830bcbe01"} Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.477318 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-7dhfs"] Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.496054 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.562156 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73f2abe5-41a2-440e-8f09-be50549a403f-logs\") pod \"horizon-675cb4b9c-cgtnr\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.562216 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73f2abe5-41a2-440e-8f09-be50549a403f-config-data\") pod \"horizon-675cb4b9c-cgtnr\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.562238 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73f2abe5-41a2-440e-8f09-be50549a403f-scripts\") pod \"horizon-675cb4b9c-cgtnr\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.562277 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55cq5\" (UniqueName: \"kubernetes.io/projected/73f2abe5-41a2-440e-8f09-be50549a403f-kube-api-access-55cq5\") pod \"horizon-675cb4b9c-cgtnr\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.562328 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73f2abe5-41a2-440e-8f09-be50549a403f-horizon-secret-key\") pod \"horizon-675cb4b9c-cgtnr\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.563076 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73f2abe5-41a2-440e-8f09-be50549a403f-logs\") pod \"horizon-675cb4b9c-cgtnr\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:23 crc kubenswrapper[4814]: W0227 16:43:23.563569 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c3924ef_1488_4784_851a_5c7c2480fa00.slice/crio-9b2a32d5645a4522e14c350624d96b515e686de135393b1c24b2cdef7887b891 WatchSource:0}: Error finding container 9b2a32d5645a4522e14c350624d96b515e686de135393b1c24b2cdef7887b891: Status 404 returned error can't find the container with id 9b2a32d5645a4522e14c350624d96b515e686de135393b1c24b2cdef7887b891 Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.565213 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73f2abe5-41a2-440e-8f09-be50549a403f-scripts\") pod \"horizon-675cb4b9c-cgtnr\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.569080 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73f2abe5-41a2-440e-8f09-be50549a403f-config-data\") pod \"horizon-675cb4b9c-cgtnr\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.599608 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73f2abe5-41a2-440e-8f09-be50549a403f-horizon-secret-key\") pod \"horizon-675cb4b9c-cgtnr\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.599619 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55cq5\" (UniqueName: \"kubernetes.io/projected/73f2abe5-41a2-440e-8f09-be50549a403f-kube-api-access-55cq5\") pod \"horizon-675cb4b9c-cgtnr\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:23 crc kubenswrapper[4814]: I0227 16:43:23.665072 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:24 crc kubenswrapper[4814]: I0227 16:43:24.313879 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-675cb4b9c-cgtnr"] Feb 27 16:43:24 crc kubenswrapper[4814]: W0227 16:43:24.328835 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73f2abe5_41a2_440e_8f09_be50549a403f.slice/crio-5da1002270283213e7941b4bbc710cc8a9e6aa1edb8b8474d8fbc9a0c6dfedba WatchSource:0}: Error finding container 5da1002270283213e7941b4bbc710cc8a9e6aa1edb8b8474d8fbc9a0c6dfedba: Status 404 returned error can't find the container with id 5da1002270283213e7941b4bbc710cc8a9e6aa1edb8b8474d8fbc9a0c6dfedba Feb 27 16:43:24 crc kubenswrapper[4814]: I0227 16:43:24.505896 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb67c60-793f-4f2c-b8e5-11fb94a857b0" path="/var/lib/kubelet/pods/4bb67c60-793f-4f2c-b8e5-11fb94a857b0/volumes" Feb 27 16:43:24 crc kubenswrapper[4814]: I0227 16:43:24.510693 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3c3924ef-1488-4784-851a-5c7c2480fa00","Type":"ContainerStarted","Data":"9b2a32d5645a4522e14c350624d96b515e686de135393b1c24b2cdef7887b891"} Feb 27 16:43:24 crc kubenswrapper[4814]: I0227 16:43:24.515426 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"80158a87-46ee-48b4-8d92-6d97c7d3bcdb","Type":"ContainerStarted","Data":"2030d99bee7765b7393b0f62914a2c348ea68090c23fb000f6540e2c5d7dd1d3"} Feb 27 16:43:24 crc kubenswrapper[4814]: I0227 16:43:24.522639 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" event={"ID":"93476b12-7c88-4276-b58f-2a6173a7eca1","Type":"ContainerStarted","Data":"3ca9f51e66aecebea6995d1f41a8b77d3b1666e0d062853132be7776f7561b7d"} Feb 27 16:43:24 crc kubenswrapper[4814]: I0227 16:43:24.522752 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:24 crc kubenswrapper[4814]: I0227 16:43:24.535026 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-675cb4b9c-cgtnr" event={"ID":"73f2abe5-41a2-440e-8f09-be50549a403f","Type":"ContainerStarted","Data":"5da1002270283213e7941b4bbc710cc8a9e6aa1edb8b8474d8fbc9a0c6dfedba"} Feb 27 16:43:24 crc kubenswrapper[4814]: I0227 16:43:24.545356 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" podStartSLOduration=4.545339017 podStartE2EDuration="4.545339017s" podCreationTimestamp="2026-02-27 16:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:43:24.539370351 +0000 UTC m=+1216.991995181" watchObservedRunningTime="2026-02-27 16:43:24.545339017 +0000 UTC m=+1216.997963847" Feb 27 16:43:25 crc kubenswrapper[4814]: I0227 16:43:25.571185 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3c3924ef-1488-4784-851a-5c7c2480fa00","Type":"ContainerStarted","Data":"c0e2764da868d4e939f9b18a6abbd8e1879216738d1cd0705ff76e8814168a53"} Feb 27 16:43:25 crc kubenswrapper[4814]: I0227 16:43:25.578160 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"80158a87-46ee-48b4-8d92-6d97c7d3bcdb","Type":"ContainerStarted","Data":"f705ac5765640a8bd3f4ce2dcd559e380cd989a1f8fedff399c17ec8f38da92e"} Feb 27 16:43:25 crc kubenswrapper[4814]: I0227 16:43:25.578268 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="80158a87-46ee-48b4-8d92-6d97c7d3bcdb" containerName="glance-log" containerID="cri-o://2030d99bee7765b7393b0f62914a2c348ea68090c23fb000f6540e2c5d7dd1d3" gracePeriod=30 Feb 27 16:43:25 crc kubenswrapper[4814]: I0227 16:43:25.578384 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="80158a87-46ee-48b4-8d92-6d97c7d3bcdb" containerName="glance-httpd" containerID="cri-o://f705ac5765640a8bd3f4ce2dcd559e380cd989a1f8fedff399c17ec8f38da92e" gracePeriod=30 Feb 27 16:43:25 crc kubenswrapper[4814]: I0227 16:43:25.613027 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.6130042190000005 podStartE2EDuration="4.613004219s" podCreationTimestamp="2026-02-27 16:43:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:43:25.606995201 +0000 UTC m=+1218.059620031" watchObservedRunningTime="2026-02-27 16:43:25.613004219 +0000 UTC m=+1218.065629049" Feb 27 16:43:26 crc kubenswrapper[4814]: I0227 16:43:26.597809 4814 generic.go:334] "Generic (PLEG): container finished" podID="80158a87-46ee-48b4-8d92-6d97c7d3bcdb" containerID="f705ac5765640a8bd3f4ce2dcd559e380cd989a1f8fedff399c17ec8f38da92e" exitCode=0 Feb 27 16:43:26 crc kubenswrapper[4814]: I0227 16:43:26.598444 4814 generic.go:334] "Generic (PLEG): container finished" podID="80158a87-46ee-48b4-8d92-6d97c7d3bcdb" containerID="2030d99bee7765b7393b0f62914a2c348ea68090c23fb000f6540e2c5d7dd1d3" exitCode=143 Feb 27 16:43:26 crc kubenswrapper[4814]: I0227 16:43:26.598533 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"80158a87-46ee-48b4-8d92-6d97c7d3bcdb","Type":"ContainerDied","Data":"f705ac5765640a8bd3f4ce2dcd559e380cd989a1f8fedff399c17ec8f38da92e"} Feb 27 16:43:26 crc kubenswrapper[4814]: I0227 16:43:26.598581 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"80158a87-46ee-48b4-8d92-6d97c7d3bcdb","Type":"ContainerDied","Data":"2030d99bee7765b7393b0f62914a2c348ea68090c23fb000f6540e2c5d7dd1d3"} Feb 27 16:43:26 crc kubenswrapper[4814]: I0227 16:43:26.604958 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3c3924ef-1488-4784-851a-5c7c2480fa00","Type":"ContainerStarted","Data":"dc548cb7d34cbc790dcb1a9abab7241efa6d0ba3b73528769736a659769b0128"} Feb 27 16:43:26 crc kubenswrapper[4814]: I0227 16:43:26.605169 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3c3924ef-1488-4784-851a-5c7c2480fa00" containerName="glance-log" containerID="cri-o://c0e2764da868d4e939f9b18a6abbd8e1879216738d1cd0705ff76e8814168a53" gracePeriod=30 Feb 27 16:43:26 crc kubenswrapper[4814]: I0227 16:43:26.606234 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3c3924ef-1488-4784-851a-5c7c2480fa00" containerName="glance-httpd" containerID="cri-o://dc548cb7d34cbc790dcb1a9abab7241efa6d0ba3b73528769736a659769b0128" gracePeriod=30 Feb 27 16:43:26 crc kubenswrapper[4814]: I0227 16:43:26.631412 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.631396385 podStartE2EDuration="5.631396385s" podCreationTimestamp="2026-02-27 16:43:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:43:26.630350413 +0000 UTC m=+1219.082975243" watchObservedRunningTime="2026-02-27 16:43:26.631396385 +0000 UTC m=+1219.084021225" Feb 27 16:43:27 crc kubenswrapper[4814]: I0227 16:43:27.624152 4814 generic.go:334] "Generic (PLEG): container finished" podID="3c3924ef-1488-4784-851a-5c7c2480fa00" containerID="c0e2764da868d4e939f9b18a6abbd8e1879216738d1cd0705ff76e8814168a53" exitCode=143 Feb 27 16:43:27 crc kubenswrapper[4814]: I0227 16:43:27.624422 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3c3924ef-1488-4784-851a-5c7c2480fa00","Type":"ContainerDied","Data":"c0e2764da868d4e939f9b18a6abbd8e1879216738d1cd0705ff76e8814168a53"} Feb 27 16:43:28 crc kubenswrapper[4814]: I0227 16:43:28.642968 4814 generic.go:334] "Generic (PLEG): container finished" podID="3c3924ef-1488-4784-851a-5c7c2480fa00" containerID="dc548cb7d34cbc790dcb1a9abab7241efa6d0ba3b73528769736a659769b0128" exitCode=0 Feb 27 16:43:28 crc kubenswrapper[4814]: I0227 16:43:28.643023 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3c3924ef-1488-4784-851a-5c7c2480fa00","Type":"ContainerDied","Data":"dc548cb7d34cbc790dcb1a9abab7241efa6d0ba3b73528769736a659769b0128"} Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.321151 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6ff5f79df7-n8ks6"] Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.338750 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-845d8f46cb-qhstj"] Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.340156 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.350684 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.376321 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-845d8f46cb-qhstj"] Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.391217 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-config-data\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.391305 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv9vm\" (UniqueName: \"kubernetes.io/projected/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-kube-api-access-bv9vm\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.391351 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-horizon-secret-key\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.391423 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-logs\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.391469 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-scripts\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.391517 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-combined-ca-bundle\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.391590 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-horizon-tls-certs\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.394664 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-675cb4b9c-cgtnr"] Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.440294 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-65784f76f6-f2pcp"] Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.441769 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.463185 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-65784f76f6-f2pcp"] Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.494019 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-logs\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.494116 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-scripts\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.494844 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-combined-ca-bundle\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.494936 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-horizon-tls-certs\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.494999 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-config-data\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.495027 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv9vm\" (UniqueName: \"kubernetes.io/projected/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-kube-api-access-bv9vm\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.495058 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-horizon-secret-key\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.495148 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-logs\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.497196 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-config-data\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.497607 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-scripts\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.501549 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-combined-ca-bundle\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.505025 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-horizon-tls-certs\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.513937 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv9vm\" (UniqueName: \"kubernetes.io/projected/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-kube-api-access-bv9vm\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.521600 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-horizon-secret-key\") pod \"horizon-845d8f46cb-qhstj\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.596443 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8af960c4-8a04-42d8-83bf-9d03c23ad333-combined-ca-bundle\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.596500 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8af960c4-8a04-42d8-83bf-9d03c23ad333-logs\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.596690 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8af960c4-8a04-42d8-83bf-9d03c23ad333-config-data\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.596788 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8af960c4-8a04-42d8-83bf-9d03c23ad333-horizon-secret-key\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.596809 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8af960c4-8a04-42d8-83bf-9d03c23ad333-scripts\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.596825 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8af960c4-8a04-42d8-83bf-9d03c23ad333-horizon-tls-certs\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.596862 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpn6m\" (UniqueName: \"kubernetes.io/projected/8af960c4-8a04-42d8-83bf-9d03c23ad333-kube-api-access-vpn6m\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.669991 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.698300 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8af960c4-8a04-42d8-83bf-9d03c23ad333-combined-ca-bundle\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.698365 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8af960c4-8a04-42d8-83bf-9d03c23ad333-logs\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.698418 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8af960c4-8a04-42d8-83bf-9d03c23ad333-config-data\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.698475 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8af960c4-8a04-42d8-83bf-9d03c23ad333-horizon-secret-key\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.698492 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8af960c4-8a04-42d8-83bf-9d03c23ad333-scripts\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.698509 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8af960c4-8a04-42d8-83bf-9d03c23ad333-horizon-tls-certs\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.698529 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpn6m\" (UniqueName: \"kubernetes.io/projected/8af960c4-8a04-42d8-83bf-9d03c23ad333-kube-api-access-vpn6m\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.700000 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8af960c4-8a04-42d8-83bf-9d03c23ad333-logs\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.700677 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8af960c4-8a04-42d8-83bf-9d03c23ad333-scripts\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.701384 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8af960c4-8a04-42d8-83bf-9d03c23ad333-config-data\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.713940 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8af960c4-8a04-42d8-83bf-9d03c23ad333-horizon-secret-key\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.714463 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8af960c4-8a04-42d8-83bf-9d03c23ad333-combined-ca-bundle\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.718274 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8af960c4-8a04-42d8-83bf-9d03c23ad333-horizon-tls-certs\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.721431 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpn6m\" (UniqueName: \"kubernetes.io/projected/8af960c4-8a04-42d8-83bf-9d03c23ad333-kube-api-access-vpn6m\") pod \"horizon-65784f76f6-f2pcp\" (UID: \"8af960c4-8a04-42d8-83bf-9d03c23ad333\") " pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:29 crc kubenswrapper[4814]: I0227 16:43:29.762637 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:30 crc kubenswrapper[4814]: I0227 16:43:30.677249 4814 generic.go:334] "Generic (PLEG): container finished" podID="1c68eb62-1f62-4ed1-a2cb-405cda573e09" containerID="ba8e4d12b3a356f51ad9bcd93b46a3983571ebb22168d59fa156bd0aa103a861" exitCode=0 Feb 27 16:43:30 crc kubenswrapper[4814]: I0227 16:43:30.677287 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bxvjh" event={"ID":"1c68eb62-1f62-4ed1-a2cb-405cda573e09","Type":"ContainerDied","Data":"ba8e4d12b3a356f51ad9bcd93b46a3983571ebb22168d59fa156bd0aa103a861"} Feb 27 16:43:31 crc kubenswrapper[4814]: I0227 16:43:31.435057 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:43:31 crc kubenswrapper[4814]: I0227 16:43:31.516296 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-qdhj8"] Feb 27 16:43:31 crc kubenswrapper[4814]: I0227 16:43:31.516615 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" podUID="6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" containerName="dnsmasq-dns" containerID="cri-o://868dca07dd0ef20b6d4dab49ac197397d1fe8947c8339c8c88f4eaf0da2a6e69" gracePeriod=10 Feb 27 16:43:31 crc kubenswrapper[4814]: I0227 16:43:31.687727 4814 generic.go:334] "Generic (PLEG): container finished" podID="6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" containerID="868dca07dd0ef20b6d4dab49ac197397d1fe8947c8339c8c88f4eaf0da2a6e69" exitCode=0 Feb 27 16:43:31 crc kubenswrapper[4814]: I0227 16:43:31.687772 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" event={"ID":"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd","Type":"ContainerDied","Data":"868dca07dd0ef20b6d4dab49ac197397d1fe8947c8339c8c88f4eaf0da2a6e69"} Feb 27 16:43:31 crc kubenswrapper[4814]: I0227 16:43:31.791356 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" podUID="6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: connect: connection refused" Feb 27 16:43:36 crc kubenswrapper[4814]: I0227 16:43:36.787347 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" podUID="6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: connect: connection refused" Feb 27 16:43:38 crc kubenswrapper[4814]: E0227 16:43:38.751164 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 27 16:43:38 crc kubenswrapper[4814]: E0227 16:43:38.751697 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n4h687h5f7h57fh688h5c5hc9h598h5ch7bh585h5fbh649h5ffh67chffh567h56ch5ffh7fh5d8h9h5f4h566h67bhc8h54h77hd9h67h66fh5c6q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-55cq5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-675cb4b9c-cgtnr_openstack(73f2abe5-41a2-440e-8f09-be50549a403f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 16:43:38 crc kubenswrapper[4814]: E0227 16:43:38.755473 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-675cb4b9c-cgtnr" podUID="73f2abe5-41a2-440e-8f09-be50549a403f" Feb 27 16:43:40 crc kubenswrapper[4814]: E0227 16:43:40.245804 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Feb 27 16:43:40 crc kubenswrapper[4814]: E0227 16:43:40.246427 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k72fh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-pq8r4_openstack(6f5812c2-9952-4125-b055-dd2888fad3f5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 16:43:40 crc kubenswrapper[4814]: E0227 16:43:40.248357 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-pq8r4" podUID="6f5812c2-9952-4125-b055-dd2888fad3f5" Feb 27 16:43:40 crc kubenswrapper[4814]: E0227 16:43:40.269005 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 27 16:43:40 crc kubenswrapper[4814]: E0227 16:43:40.269210 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nc6h659h5bfh7fh696h9ch656h544h654h57h5b6h669h696h667h657h5c6h58dh68h87hf8h696h656h7dh59bh544h654h54dhchf5h54dh574hc5q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mk6b4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6ff5f79df7-n8ks6_openstack(548a8aff-dc75-4e9f-857a-0acab993607c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 16:43:40 crc kubenswrapper[4814]: E0227 16:43:40.282954 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-6ff5f79df7-n8ks6" podUID="548a8aff-dc75-4e9f-857a-0acab993607c" Feb 27 16:43:40 crc kubenswrapper[4814]: E0227 16:43:40.328082 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 27 16:43:40 crc kubenswrapper[4814]: E0227 16:43:40.328328 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfch5bdh577h648h545h567hf4h5bch86h74h95h5dchd7h78hbh9bhf4h59bh8h5cbhbh5bh56ch567h665h66bh54h7ch664h64h68fhd9q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k9lgb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5c887c6797-kst67_openstack(3a57db6d-9660-4c37-847d-b25c6d65e147): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 16:43:40 crc kubenswrapper[4814]: E0227 16:43:40.330797 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5c887c6797-kst67" podUID="3a57db6d-9660-4c37-847d-b25c6d65e147" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.383695 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.532359 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9cj4\" (UniqueName: \"kubernetes.io/projected/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-kube-api-access-k9cj4\") pod \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.532939 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.536873 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-config-data\") pod \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.536928 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-logs\") pod \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.536959 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-httpd-run\") pod \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.537015 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-public-tls-certs\") pod \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.537397 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-combined-ca-bundle\") pod \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.537472 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-scripts\") pod \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\" (UID: \"80158a87-46ee-48b4-8d92-6d97c7d3bcdb\") " Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.537658 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "80158a87-46ee-48b4-8d92-6d97c7d3bcdb" (UID: "80158a87-46ee-48b4-8d92-6d97c7d3bcdb"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.537681 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-logs" (OuterVolumeSpecName: "logs") pod "80158a87-46ee-48b4-8d92-6d97c7d3bcdb" (UID: "80158a87-46ee-48b4-8d92-6d97c7d3bcdb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.538557 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.538609 4814 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.553456 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "80158a87-46ee-48b4-8d92-6d97c7d3bcdb" (UID: "80158a87-46ee-48b4-8d92-6d97c7d3bcdb"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.553473 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-kube-api-access-k9cj4" (OuterVolumeSpecName: "kube-api-access-k9cj4") pod "80158a87-46ee-48b4-8d92-6d97c7d3bcdb" (UID: "80158a87-46ee-48b4-8d92-6d97c7d3bcdb"). InnerVolumeSpecName "kube-api-access-k9cj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.566022 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80158a87-46ee-48b4-8d92-6d97c7d3bcdb" (UID: "80158a87-46ee-48b4-8d92-6d97c7d3bcdb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.569296 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-scripts" (OuterVolumeSpecName: "scripts") pod "80158a87-46ee-48b4-8d92-6d97c7d3bcdb" (UID: "80158a87-46ee-48b4-8d92-6d97c7d3bcdb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.609364 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "80158a87-46ee-48b4-8d92-6d97c7d3bcdb" (UID: "80158a87-46ee-48b4-8d92-6d97c7d3bcdb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.640045 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.640079 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.640089 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9cj4\" (UniqueName: \"kubernetes.io/projected/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-kube-api-access-k9cj4\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.640117 4814 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.640135 4814 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.667393 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-config-data" (OuterVolumeSpecName: "config-data") pod "80158a87-46ee-48b4-8d92-6d97c7d3bcdb" (UID: "80158a87-46ee-48b4-8d92-6d97c7d3bcdb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.726979 4814 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.747475 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80158a87-46ee-48b4-8d92-6d97c7d3bcdb-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.747519 4814 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.808280 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.809057 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"80158a87-46ee-48b4-8d92-6d97c7d3bcdb","Type":"ContainerDied","Data":"d9711b917b1cf3291855dcb95258f508091299b101788ad3ce6a07e9e97bef97"} Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.809115 4814 scope.go:117] "RemoveContainer" containerID="f705ac5765640a8bd3f4ce2dcd559e380cd989a1f8fedff399c17ec8f38da92e" Feb 27 16:43:40 crc kubenswrapper[4814]: E0227 16:43:40.816936 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-pq8r4" podUID="6f5812c2-9952-4125-b055-dd2888fad3f5" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.925478 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.932584 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.953745 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 16:43:40 crc kubenswrapper[4814]: E0227 16:43:40.954279 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80158a87-46ee-48b4-8d92-6d97c7d3bcdb" containerName="glance-httpd" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.954353 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="80158a87-46ee-48b4-8d92-6d97c7d3bcdb" containerName="glance-httpd" Feb 27 16:43:40 crc kubenswrapper[4814]: E0227 16:43:40.954454 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80158a87-46ee-48b4-8d92-6d97c7d3bcdb" containerName="glance-log" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.954508 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="80158a87-46ee-48b4-8d92-6d97c7d3bcdb" containerName="glance-log" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.954711 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="80158a87-46ee-48b4-8d92-6d97c7d3bcdb" containerName="glance-log" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.954768 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="80158a87-46ee-48b4-8d92-6d97c7d3bcdb" containerName="glance-httpd" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.955869 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.958655 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.958841 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 27 16:43:40 crc kubenswrapper[4814]: I0227 16:43:40.981570 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.156147 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.156270 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-config-data\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.156297 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.156318 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvmzz\" (UniqueName: \"kubernetes.io/projected/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-kube-api-access-gvmzz\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.156360 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-scripts\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.156382 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.156399 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-logs\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.156434 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.257614 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.257680 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-config-data\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.257706 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.257724 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvmzz\" (UniqueName: \"kubernetes.io/projected/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-kube-api-access-gvmzz\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.257766 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-scripts\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.257792 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.257813 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-logs\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.257843 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.258085 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.259333 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.263688 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-logs\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.264305 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-scripts\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.264949 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.265046 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-config-data\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.265315 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.283518 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvmzz\" (UniqueName: \"kubernetes.io/projected/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-kube-api-access-gvmzz\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.305246 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " pod="openstack/glance-default-external-api-0" Feb 27 16:43:41 crc kubenswrapper[4814]: E0227 16:43:41.400714 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Feb 27 16:43:41 crc kubenswrapper[4814]: E0227 16:43:41.401139 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n7dh66bh85h5cbh655h58dh596h685h77h697hddh654h59ch5dchf7h5bdh665h5d6h68fh648h59dh6bh68bh5d7h677h88h5bfhbh69h694h547h564q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fdqh8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(01245f0c-38fe-4f9a-93e5-c5c5d5f32259): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 16:43:41 crc kubenswrapper[4814]: I0227 16:43:41.575291 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 16:43:42 crc kubenswrapper[4814]: I0227 16:43:42.497057 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80158a87-46ee-48b4-8d92-6d97c7d3bcdb" path="/var/lib/kubelet/pods/80158a87-46ee-48b4-8d92-6d97c7d3bcdb/volumes" Feb 27 16:43:46 crc kubenswrapper[4814]: I0227 16:43:46.787937 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" podUID="6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: i/o timeout" Feb 27 16:43:46 crc kubenswrapper[4814]: I0227 16:43:46.788841 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:49 crc kubenswrapper[4814]: I0227 16:43:49.919834 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-675cb4b9c-cgtnr" event={"ID":"73f2abe5-41a2-440e-8f09-be50549a403f","Type":"ContainerDied","Data":"5da1002270283213e7941b4bbc710cc8a9e6aa1edb8b8474d8fbc9a0c6dfedba"} Feb 27 16:43:49 crc kubenswrapper[4814]: I0227 16:43:49.920382 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5da1002270283213e7941b4bbc710cc8a9e6aa1edb8b8474d8fbc9a0c6dfedba" Feb 27 16:43:49 crc kubenswrapper[4814]: I0227 16:43:49.923646 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bxvjh" event={"ID":"1c68eb62-1f62-4ed1-a2cb-405cda573e09","Type":"ContainerDied","Data":"4234d0f926810b9d3c48e7263c539f370a012a2c0d502aa51105aa1af288c8dd"} Feb 27 16:43:49 crc kubenswrapper[4814]: I0227 16:43:49.923707 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4234d0f926810b9d3c48e7263c539f370a012a2c0d502aa51105aa1af288c8dd" Feb 27 16:43:49 crc kubenswrapper[4814]: I0227 16:43:49.926848 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" event={"ID":"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd","Type":"ContainerDied","Data":"2f7e226474e68967bee396e70bed60a819c2b72b5259d688cd8f548731ab7798"} Feb 27 16:43:49 crc kubenswrapper[4814]: I0227 16:43:49.926883 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f7e226474e68967bee396e70bed60a819c2b72b5259d688cd8f548731ab7798" Feb 27 16:43:49 crc kubenswrapper[4814]: I0227 16:43:49.928655 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c887c6797-kst67" event={"ID":"3a57db6d-9660-4c37-847d-b25c6d65e147","Type":"ContainerDied","Data":"0a03f5372e573e1f531baf20c83f62cb403192409cb87766fb4412843edf71c3"} Feb 27 16:43:49 crc kubenswrapper[4814]: I0227 16:43:49.928823 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a03f5372e573e1f531baf20c83f62cb403192409cb87766fb4412843edf71c3" Feb 27 16:43:49 crc kubenswrapper[4814]: I0227 16:43:49.932071 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3c3924ef-1488-4784-851a-5c7c2480fa00","Type":"ContainerDied","Data":"9b2a32d5645a4522e14c350624d96b515e686de135393b1c24b2cdef7887b891"} Feb 27 16:43:49 crc kubenswrapper[4814]: I0227 16:43:49.932115 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b2a32d5645a4522e14c350624d96b515e686de135393b1c24b2cdef7887b891" Feb 27 16:43:49 crc kubenswrapper[4814]: I0227 16:43:49.933829 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ff5f79df7-n8ks6" event={"ID":"548a8aff-dc75-4e9f-857a-0acab993607c","Type":"ContainerDied","Data":"e994f8368bf85075374a41b11ee08168333adf6ee2e00a348c3a983830bcbe01"} Feb 27 16:43:49 crc kubenswrapper[4814]: I0227 16:43:49.934594 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e994f8368bf85075374a41b11ee08168333adf6ee2e00a348c3a983830bcbe01" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.026943 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.031689 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.039573 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.046054 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.056385 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.067584 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.135413 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73f2abe5-41a2-440e-8f09-be50549a403f-config-data\") pod \"73f2abe5-41a2-440e-8f09-be50549a403f\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.135464 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-config-data\") pod \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.135518 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73f2abe5-41a2-440e-8f09-be50549a403f-scripts\") pod \"73f2abe5-41a2-440e-8f09-be50549a403f\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.135546 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-ovsdbserver-sb\") pod \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.136172 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73f2abe5-41a2-440e-8f09-be50549a403f-scripts" (OuterVolumeSpecName: "scripts") pod "73f2abe5-41a2-440e-8f09-be50549a403f" (UID: "73f2abe5-41a2-440e-8f09-be50549a403f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.136245 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp558\" (UniqueName: \"kubernetes.io/projected/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-kube-api-access-hp558\") pod \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.136655 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-ovsdbserver-nb\") pod \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.136683 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-dns-svc\") pod \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.136759 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-combined-ca-bundle\") pod \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.136775 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-fernet-keys\") pod \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.136795 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-scripts\") pod \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.136824 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jm9x\" (UniqueName: \"kubernetes.io/projected/1c68eb62-1f62-4ed1-a2cb-405cda573e09-kube-api-access-7jm9x\") pod \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.136854 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-config\") pod \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.136901 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73f2abe5-41a2-440e-8f09-be50549a403f-horizon-secret-key\") pod \"73f2abe5-41a2-440e-8f09-be50549a403f\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.136926 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-dns-swift-storage-0\") pod \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\" (UID: \"6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.136960 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-credential-keys\") pod \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\" (UID: \"1c68eb62-1f62-4ed1-a2cb-405cda573e09\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.136982 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55cq5\" (UniqueName: \"kubernetes.io/projected/73f2abe5-41a2-440e-8f09-be50549a403f-kube-api-access-55cq5\") pod \"73f2abe5-41a2-440e-8f09-be50549a403f\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.137008 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73f2abe5-41a2-440e-8f09-be50549a403f-logs\") pod \"73f2abe5-41a2-440e-8f09-be50549a403f\" (UID: \"73f2abe5-41a2-440e-8f09-be50549a403f\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.137513 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73f2abe5-41a2-440e-8f09-be50549a403f-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.137873 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73f2abe5-41a2-440e-8f09-be50549a403f-config-data" (OuterVolumeSpecName: "config-data") pod "73f2abe5-41a2-440e-8f09-be50549a403f" (UID: "73f2abe5-41a2-440e-8f09-be50549a403f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.144406 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-kube-api-access-hp558" (OuterVolumeSpecName: "kube-api-access-hp558") pod "6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" (UID: "6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd"). InnerVolumeSpecName "kube-api-access-hp558". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.149744 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c68eb62-1f62-4ed1-a2cb-405cda573e09-kube-api-access-7jm9x" (OuterVolumeSpecName: "kube-api-access-7jm9x") pod "1c68eb62-1f62-4ed1-a2cb-405cda573e09" (UID: "1c68eb62-1f62-4ed1-a2cb-405cda573e09"). InnerVolumeSpecName "kube-api-access-7jm9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.151754 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73f2abe5-41a2-440e-8f09-be50549a403f-logs" (OuterVolumeSpecName: "logs") pod "73f2abe5-41a2-440e-8f09-be50549a403f" (UID: "73f2abe5-41a2-440e-8f09-be50549a403f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.156507 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73f2abe5-41a2-440e-8f09-be50549a403f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "73f2abe5-41a2-440e-8f09-be50549a403f" (UID: "73f2abe5-41a2-440e-8f09-be50549a403f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.157426 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1c68eb62-1f62-4ed1-a2cb-405cda573e09" (UID: "1c68eb62-1f62-4ed1-a2cb-405cda573e09"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.158230 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-scripts" (OuterVolumeSpecName: "scripts") pod "1c68eb62-1f62-4ed1-a2cb-405cda573e09" (UID: "1c68eb62-1f62-4ed1-a2cb-405cda573e09"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.158297 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1c68eb62-1f62-4ed1-a2cb-405cda573e09" (UID: "1c68eb62-1f62-4ed1-a2cb-405cda573e09"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.178379 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73f2abe5-41a2-440e-8f09-be50549a403f-kube-api-access-55cq5" (OuterVolumeSpecName: "kube-api-access-55cq5") pod "73f2abe5-41a2-440e-8f09-be50549a403f" (UID: "73f2abe5-41a2-440e-8f09-be50549a403f"). InnerVolumeSpecName "kube-api-access-55cq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.181925 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-config-data" (OuterVolumeSpecName: "config-data") pod "1c68eb62-1f62-4ed1-a2cb-405cda573e09" (UID: "1c68eb62-1f62-4ed1-a2cb-405cda573e09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.213789 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1c68eb62-1f62-4ed1-a2cb-405cda573e09" (UID: "1c68eb62-1f62-4ed1-a2cb-405cda573e09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.217804 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" (UID: "6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.234021 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" (UID: "6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.234974 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" (UID: "6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.237143 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" (UID: "6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239244 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-internal-tls-certs\") pod \"3c3924ef-1488-4784-851a-5c7c2480fa00\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239304 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mk6b4\" (UniqueName: \"kubernetes.io/projected/548a8aff-dc75-4e9f-857a-0acab993607c-kube-api-access-mk6b4\") pod \"548a8aff-dc75-4e9f-857a-0acab993607c\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239335 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c3924ef-1488-4784-851a-5c7c2480fa00-logs\") pod \"3c3924ef-1488-4784-851a-5c7c2480fa00\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239359 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/548a8aff-dc75-4e9f-857a-0acab993607c-scripts\") pod \"548a8aff-dc75-4e9f-857a-0acab993607c\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239412 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-scripts\") pod \"3c3924ef-1488-4784-851a-5c7c2480fa00\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239436 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a57db6d-9660-4c37-847d-b25c6d65e147-logs\") pod \"3a57db6d-9660-4c37-847d-b25c6d65e147\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239467 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9lgb\" (UniqueName: \"kubernetes.io/projected/3a57db6d-9660-4c37-847d-b25c6d65e147-kube-api-access-k9lgb\") pod \"3a57db6d-9660-4c37-847d-b25c6d65e147\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239491 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-combined-ca-bundle\") pod \"3c3924ef-1488-4784-851a-5c7c2480fa00\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239521 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/548a8aff-dc75-4e9f-857a-0acab993607c-config-data\") pod \"548a8aff-dc75-4e9f-857a-0acab993607c\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239549 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bnjq\" (UniqueName: \"kubernetes.io/projected/3c3924ef-1488-4784-851a-5c7c2480fa00-kube-api-access-6bnjq\") pod \"3c3924ef-1488-4784-851a-5c7c2480fa00\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239575 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3c3924ef-1488-4784-851a-5c7c2480fa00-httpd-run\") pod \"3c3924ef-1488-4784-851a-5c7c2480fa00\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239605 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a57db6d-9660-4c37-847d-b25c6d65e147-config-data\") pod \"3a57db6d-9660-4c37-847d-b25c6d65e147\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239627 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-config-data\") pod \"3c3924ef-1488-4784-851a-5c7c2480fa00\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239648 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/548a8aff-dc75-4e9f-857a-0acab993607c-logs\") pod \"548a8aff-dc75-4e9f-857a-0acab993607c\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239726 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3a57db6d-9660-4c37-847d-b25c6d65e147-horizon-secret-key\") pod \"3a57db6d-9660-4c37-847d-b25c6d65e147\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239743 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"3c3924ef-1488-4784-851a-5c7c2480fa00\" (UID: \"3c3924ef-1488-4784-851a-5c7c2480fa00\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239793 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/548a8aff-dc75-4e9f-857a-0acab993607c-horizon-secret-key\") pod \"548a8aff-dc75-4e9f-857a-0acab993607c\" (UID: \"548a8aff-dc75-4e9f-857a-0acab993607c\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.239817 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a57db6d-9660-4c37-847d-b25c6d65e147-scripts\") pod \"3a57db6d-9660-4c37-847d-b25c6d65e147\" (UID: \"3a57db6d-9660-4c37-847d-b25c6d65e147\") " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240168 4814 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240179 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240189 4814 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240198 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240206 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jm9x\" (UniqueName: \"kubernetes.io/projected/1c68eb62-1f62-4ed1-a2cb-405cda573e09-kube-api-access-7jm9x\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240216 4814 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73f2abe5-41a2-440e-8f09-be50549a403f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240225 4814 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240233 4814 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240242 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55cq5\" (UniqueName: \"kubernetes.io/projected/73f2abe5-41a2-440e-8f09-be50549a403f-kube-api-access-55cq5\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240269 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73f2abe5-41a2-440e-8f09-be50549a403f-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240281 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c68eb62-1f62-4ed1-a2cb-405cda573e09-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240292 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73f2abe5-41a2-440e-8f09-be50549a403f-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240303 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240313 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp558\" (UniqueName: \"kubernetes.io/projected/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-kube-api-access-hp558\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240323 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240337 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c3924ef-1488-4784-851a-5c7c2480fa00-logs" (OuterVolumeSpecName: "logs") pod "3c3924ef-1488-4784-851a-5c7c2480fa00" (UID: "3c3924ef-1488-4784-851a-5c7c2480fa00"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.240745 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a57db6d-9660-4c37-847d-b25c6d65e147-scripts" (OuterVolumeSpecName: "scripts") pod "3a57db6d-9660-4c37-847d-b25c6d65e147" (UID: "3a57db6d-9660-4c37-847d-b25c6d65e147"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.241344 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a57db6d-9660-4c37-847d-b25c6d65e147-logs" (OuterVolumeSpecName: "logs") pod "3a57db6d-9660-4c37-847d-b25c6d65e147" (UID: "3a57db6d-9660-4c37-847d-b25c6d65e147"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.242546 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c3924ef-1488-4784-851a-5c7c2480fa00-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3c3924ef-1488-4784-851a-5c7c2480fa00" (UID: "3c3924ef-1488-4784-851a-5c7c2480fa00"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.244314 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c3924ef-1488-4784-851a-5c7c2480fa00-kube-api-access-6bnjq" (OuterVolumeSpecName: "kube-api-access-6bnjq") pod "3c3924ef-1488-4784-851a-5c7c2480fa00" (UID: "3c3924ef-1488-4784-851a-5c7c2480fa00"). InnerVolumeSpecName "kube-api-access-6bnjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.244865 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-scripts" (OuterVolumeSpecName: "scripts") pod "3c3924ef-1488-4784-851a-5c7c2480fa00" (UID: "3c3924ef-1488-4784-851a-5c7c2480fa00"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.244916 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a57db6d-9660-4c37-847d-b25c6d65e147-kube-api-access-k9lgb" (OuterVolumeSpecName: "kube-api-access-k9lgb") pod "3a57db6d-9660-4c37-847d-b25c6d65e147" (UID: "3a57db6d-9660-4c37-847d-b25c6d65e147"). InnerVolumeSpecName "kube-api-access-k9lgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.245548 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/548a8aff-dc75-4e9f-857a-0acab993607c-logs" (OuterVolumeSpecName: "logs") pod "548a8aff-dc75-4e9f-857a-0acab993607c" (UID: "548a8aff-dc75-4e9f-857a-0acab993607c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.245982 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/548a8aff-dc75-4e9f-857a-0acab993607c-scripts" (OuterVolumeSpecName: "scripts") pod "548a8aff-dc75-4e9f-857a-0acab993607c" (UID: "548a8aff-dc75-4e9f-857a-0acab993607c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.246242 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/548a8aff-dc75-4e9f-857a-0acab993607c-config-data" (OuterVolumeSpecName: "config-data") pod "548a8aff-dc75-4e9f-857a-0acab993607c" (UID: "548a8aff-dc75-4e9f-857a-0acab993607c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.246832 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a57db6d-9660-4c37-847d-b25c6d65e147-config-data" (OuterVolumeSpecName: "config-data") pod "3a57db6d-9660-4c37-847d-b25c6d65e147" (UID: "3a57db6d-9660-4c37-847d-b25c6d65e147"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.247040 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "3c3924ef-1488-4784-851a-5c7c2480fa00" (UID: "3c3924ef-1488-4784-851a-5c7c2480fa00"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.247509 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a57db6d-9660-4c37-847d-b25c6d65e147-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "3a57db6d-9660-4c37-847d-b25c6d65e147" (UID: "3a57db6d-9660-4c37-847d-b25c6d65e147"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.249317 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/548a8aff-dc75-4e9f-857a-0acab993607c-kube-api-access-mk6b4" (OuterVolumeSpecName: "kube-api-access-mk6b4") pod "548a8aff-dc75-4e9f-857a-0acab993607c" (UID: "548a8aff-dc75-4e9f-857a-0acab993607c"). InnerVolumeSpecName "kube-api-access-mk6b4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.250119 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/548a8aff-dc75-4e9f-857a-0acab993607c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "548a8aff-dc75-4e9f-857a-0acab993607c" (UID: "548a8aff-dc75-4e9f-857a-0acab993607c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.257322 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-config" (OuterVolumeSpecName: "config") pod "6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" (UID: "6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.269978 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c3924ef-1488-4784-851a-5c7c2480fa00" (UID: "3c3924ef-1488-4784-851a-5c7c2480fa00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.290288 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3c3924ef-1488-4784-851a-5c7c2480fa00" (UID: "3c3924ef-1488-4784-851a-5c7c2480fa00"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.294945 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-config-data" (OuterVolumeSpecName: "config-data") pod "3c3924ef-1488-4784-851a-5c7c2480fa00" (UID: "3c3924ef-1488-4784-851a-5c7c2480fa00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.343962 4814 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.344346 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mk6b4\" (UniqueName: \"kubernetes.io/projected/548a8aff-dc75-4e9f-857a-0acab993607c-kube-api-access-mk6b4\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.344412 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c3924ef-1488-4784-851a-5c7c2480fa00-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.344466 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/548a8aff-dc75-4e9f-857a-0acab993607c-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.344524 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.344587 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.344642 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a57db6d-9660-4c37-847d-b25c6d65e147-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.344698 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9lgb\" (UniqueName: \"kubernetes.io/projected/3a57db6d-9660-4c37-847d-b25c6d65e147-kube-api-access-k9lgb\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.344747 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.344801 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/548a8aff-dc75-4e9f-857a-0acab993607c-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.344855 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bnjq\" (UniqueName: \"kubernetes.io/projected/3c3924ef-1488-4784-851a-5c7c2480fa00-kube-api-access-6bnjq\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.344911 4814 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3c3924ef-1488-4784-851a-5c7c2480fa00-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.344963 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a57db6d-9660-4c37-847d-b25c6d65e147-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.345022 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c3924ef-1488-4784-851a-5c7c2480fa00-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.345075 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/548a8aff-dc75-4e9f-857a-0acab993607c-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.345123 4814 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3a57db6d-9660-4c37-847d-b25c6d65e147-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.345235 4814 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.345324 4814 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/548a8aff-dc75-4e9f-857a-0acab993607c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.345385 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a57db6d-9660-4c37-847d-b25c6d65e147-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.364951 4814 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.447162 4814 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.943131 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c887c6797-kst67" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.943164 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.943164 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.943182 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-675cb4b9c-cgtnr" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.943182 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bxvjh" Feb 27 16:43:50 crc kubenswrapper[4814]: I0227 16:43:50.943170 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ff5f79df7-n8ks6" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.004594 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-675cb4b9c-cgtnr"] Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.009827 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-675cb4b9c-cgtnr"] Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.033743 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6ff5f79df7-n8ks6"] Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.044207 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6ff5f79df7-n8ks6"] Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.051786 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-qdhj8"] Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.069373 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-qdhj8"] Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.098278 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5c887c6797-kst67"] Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.122377 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5c887c6797-kst67"] Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.130921 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.145187 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.152203 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 16:43:51 crc kubenswrapper[4814]: E0227 16:43:51.152587 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" containerName="init" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.152600 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" containerName="init" Feb 27 16:43:51 crc kubenswrapper[4814]: E0227 16:43:51.152610 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" containerName="dnsmasq-dns" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.152616 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" containerName="dnsmasq-dns" Feb 27 16:43:51 crc kubenswrapper[4814]: E0227 16:43:51.152631 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3924ef-1488-4784-851a-5c7c2480fa00" containerName="glance-log" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.152637 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3924ef-1488-4784-851a-5c7c2480fa00" containerName="glance-log" Feb 27 16:43:51 crc kubenswrapper[4814]: E0227 16:43:51.152654 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3924ef-1488-4784-851a-5c7c2480fa00" containerName="glance-httpd" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.152660 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3924ef-1488-4784-851a-5c7c2480fa00" containerName="glance-httpd" Feb 27 16:43:51 crc kubenswrapper[4814]: E0227 16:43:51.152686 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c68eb62-1f62-4ed1-a2cb-405cda573e09" containerName="keystone-bootstrap" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.152693 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c68eb62-1f62-4ed1-a2cb-405cda573e09" containerName="keystone-bootstrap" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.152846 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c68eb62-1f62-4ed1-a2cb-405cda573e09" containerName="keystone-bootstrap" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.152859 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c3924ef-1488-4784-851a-5c7c2480fa00" containerName="glance-log" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.152865 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" containerName="dnsmasq-dns" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.152873 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c3924ef-1488-4784-851a-5c7c2480fa00" containerName="glance-httpd" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.154455 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.158717 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-bxvjh"] Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.158897 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.158996 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.164691 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-bxvjh"] Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.170847 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.225089 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-88h8r"] Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.226931 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.229219 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.229635 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.229816 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.229924 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.230109 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-m5bsz" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.237604 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-88h8r"] Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.261725 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.261787 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-logs\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.261809 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdqtd\" (UniqueName: \"kubernetes.io/projected/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-kube-api-access-zdqtd\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.261827 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.262034 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.262274 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.262327 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.262398 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.364115 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.364181 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-credential-keys\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.364207 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.364222 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.364272 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.364310 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-fernet-keys\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.364325 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-combined-ca-bundle\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.364360 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.364390 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.364407 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-logs\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.364422 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdqtd\" (UniqueName: \"kubernetes.io/projected/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-kube-api-access-zdqtd\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.364439 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-scripts\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.364454 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-config-data\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.364474 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v94rk\" (UniqueName: \"kubernetes.io/projected/6f56fd06-b678-48f2-9f0d-0234a53b528b-kube-api-access-v94rk\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.365419 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.365619 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.366360 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-logs\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.369878 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.374109 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.374726 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.375173 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.380918 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdqtd\" (UniqueName: \"kubernetes.io/projected/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-kube-api-access-zdqtd\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.399822 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.465760 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-credential-keys\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.466098 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-fernet-keys\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.466141 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-combined-ca-bundle\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.466315 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-scripts\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.466356 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-config-data\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.468084 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v94rk\" (UniqueName: \"kubernetes.io/projected/6f56fd06-b678-48f2-9f0d-0234a53b528b-kube-api-access-v94rk\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.471397 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-config-data\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.471504 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-combined-ca-bundle\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.472841 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-credential-keys\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.475526 4814 scope.go:117] "RemoveContainer" containerID="2030d99bee7765b7393b0f62914a2c348ea68090c23fb000f6540e2c5d7dd1d3" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.475603 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-scripts\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.481396 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-fernet-keys\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.485028 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v94rk\" (UniqueName: \"kubernetes.io/projected/6f56fd06-b678-48f2-9f0d-0234a53b528b-kube-api-access-v94rk\") pod \"keystone-bootstrap-88h8r\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.485037 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 16:43:51 crc kubenswrapper[4814]: E0227 16:43:51.506694 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Feb 27 16:43:51 crc kubenswrapper[4814]: E0227 16:43:51.506839 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rpmlb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-vqkl5_openstack(7bcf9bc6-1fb4-47b4-b12d-017902e4992e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 16:43:51 crc kubenswrapper[4814]: E0227 16:43:51.508694 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-vqkl5" podUID="7bcf9bc6-1fb4-47b4-b12d-017902e4992e" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.551121 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.789205 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7ff5475cc9-qdhj8" podUID="6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: i/o timeout" Feb 27 16:43:51 crc kubenswrapper[4814]: I0227 16:43:51.882619 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-845d8f46cb-qhstj"] Feb 27 16:43:51 crc kubenswrapper[4814]: W0227 16:43:51.949227 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c9011a2_17bc_42ba_a6cf_13ec98cd75b1.slice/crio-8490be3507800d70feb1b18bf4bbe755740d3013e1f131ad3df196e1c73cec7b WatchSource:0}: Error finding container 8490be3507800d70feb1b18bf4bbe755740d3013e1f131ad3df196e1c73cec7b: Status 404 returned error can't find the container with id 8490be3507800d70feb1b18bf4bbe755740d3013e1f131ad3df196e1c73cec7b Feb 27 16:43:52 crc kubenswrapper[4814]: E0227 16:43:52.005928 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-vqkl5" podUID="7bcf9bc6-1fb4-47b4-b12d-017902e4992e" Feb 27 16:43:52 crc kubenswrapper[4814]: I0227 16:43:52.100384 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-65784f76f6-f2pcp"] Feb 27 16:43:52 crc kubenswrapper[4814]: I0227 16:43:52.427133 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 16:43:52 crc kubenswrapper[4814]: W0227 16:43:52.432080 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0886f8c2_6ec0_46af_8ebe_2e2cdd853d31.slice/crio-93513a28d90cb7ec06a4af69840230c2530d021ffcb4ad375e1bd50fa030547f WatchSource:0}: Error finding container 93513a28d90cb7ec06a4af69840230c2530d021ffcb4ad375e1bd50fa030547f: Status 404 returned error can't find the container with id 93513a28d90cb7ec06a4af69840230c2530d021ffcb4ad375e1bd50fa030547f Feb 27 16:43:52 crc kubenswrapper[4814]: I0227 16:43:52.506617 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c68eb62-1f62-4ed1-a2cb-405cda573e09" path="/var/lib/kubelet/pods/1c68eb62-1f62-4ed1-a2cb-405cda573e09/volumes" Feb 27 16:43:52 crc kubenswrapper[4814]: I0227 16:43:52.507550 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a57db6d-9660-4c37-847d-b25c6d65e147" path="/var/lib/kubelet/pods/3a57db6d-9660-4c37-847d-b25c6d65e147/volumes" Feb 27 16:43:52 crc kubenswrapper[4814]: I0227 16:43:52.508074 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c3924ef-1488-4784-851a-5c7c2480fa00" path="/var/lib/kubelet/pods/3c3924ef-1488-4784-851a-5c7c2480fa00/volumes" Feb 27 16:43:52 crc kubenswrapper[4814]: W0227 16:43:52.509910 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f56fd06_b678_48f2_9f0d_0234a53b528b.slice/crio-1a23ac6768bf39d8b065376f92384a96dac2703a6e8825bb53b875bcc595900d WatchSource:0}: Error finding container 1a23ac6768bf39d8b065376f92384a96dac2703a6e8825bb53b875bcc595900d: Status 404 returned error can't find the container with id 1a23ac6768bf39d8b065376f92384a96dac2703a6e8825bb53b875bcc595900d Feb 27 16:43:52 crc kubenswrapper[4814]: I0227 16:43:52.513365 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="548a8aff-dc75-4e9f-857a-0acab993607c" path="/var/lib/kubelet/pods/548a8aff-dc75-4e9f-857a-0acab993607c/volumes" Feb 27 16:43:52 crc kubenswrapper[4814]: I0227 16:43:52.513868 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd" path="/var/lib/kubelet/pods/6b9ef32e-99c7-46f7-8b1c-5e9cb53fdbfd/volumes" Feb 27 16:43:52 crc kubenswrapper[4814]: I0227 16:43:52.514709 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73f2abe5-41a2-440e-8f09-be50549a403f" path="/var/lib/kubelet/pods/73f2abe5-41a2-440e-8f09-be50549a403f/volumes" Feb 27 16:43:52 crc kubenswrapper[4814]: I0227 16:43:52.515135 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-88h8r"] Feb 27 16:43:52 crc kubenswrapper[4814]: I0227 16:43:52.748487 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 16:43:52 crc kubenswrapper[4814]: W0227 16:43:52.755239 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f8e8e66_aaf0_43d6_b990_e046d17f2fb5.slice/crio-82c9148f20c6fdb72bfaae8d94beaf97dd4c55f9689a736c2fcd251db6504b97 WatchSource:0}: Error finding container 82c9148f20c6fdb72bfaae8d94beaf97dd4c55f9689a736c2fcd251db6504b97: Status 404 returned error can't find the container with id 82c9148f20c6fdb72bfaae8d94beaf97dd4c55f9689a736c2fcd251db6504b97 Feb 27 16:43:53 crc kubenswrapper[4814]: I0227 16:43:53.016745 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-845d8f46cb-qhstj" event={"ID":"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1","Type":"ContainerStarted","Data":"8490be3507800d70feb1b18bf4bbe755740d3013e1f131ad3df196e1c73cec7b"} Feb 27 16:43:53 crc kubenswrapper[4814]: I0227 16:43:53.018990 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31","Type":"ContainerStarted","Data":"93513a28d90cb7ec06a4af69840230c2530d021ffcb4ad375e1bd50fa030547f"} Feb 27 16:43:53 crc kubenswrapper[4814]: I0227 16:43:53.044987 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-88h8r" event={"ID":"6f56fd06-b678-48f2-9f0d-0234a53b528b","Type":"ContainerStarted","Data":"67fab44cb9ddab92dbc952ba5825863ced0e115a1f4e9dac622240cd0f1e0176"} Feb 27 16:43:53 crc kubenswrapper[4814]: I0227 16:43:53.045091 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-88h8r" event={"ID":"6f56fd06-b678-48f2-9f0d-0234a53b528b","Type":"ContainerStarted","Data":"1a23ac6768bf39d8b065376f92384a96dac2703a6e8825bb53b875bcc595900d"} Feb 27 16:43:53 crc kubenswrapper[4814]: I0227 16:43:53.049572 4814 generic.go:334] "Generic (PLEG): container finished" podID="5e413768-6d50-4725-a809-3658e913cfed" containerID="0615c5bc9eee0c9e2440a32cd8f263d45418c571909c1a6d783625924de3e53c" exitCode=0 Feb 27 16:43:53 crc kubenswrapper[4814]: I0227 16:43:53.049646 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-th62b" event={"ID":"5e413768-6d50-4725-a809-3658e913cfed","Type":"ContainerDied","Data":"0615c5bc9eee0c9e2440a32cd8f263d45418c571909c1a6d783625924de3e53c"} Feb 27 16:43:53 crc kubenswrapper[4814]: I0227 16:43:53.052029 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65784f76f6-f2pcp" event={"ID":"8af960c4-8a04-42d8-83bf-9d03c23ad333","Type":"ContainerStarted","Data":"dec509f4090490e80f852609ba7162241dcc2a89c7661c4bcb4896e531d96f9b"} Feb 27 16:43:53 crc kubenswrapper[4814]: I0227 16:43:53.052051 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65784f76f6-f2pcp" event={"ID":"8af960c4-8a04-42d8-83bf-9d03c23ad333","Type":"ContainerStarted","Data":"55c234ea8a9ec40624eaa4d15f127942e19be6be9a0cd7c7791609ad2e930367"} Feb 27 16:43:53 crc kubenswrapper[4814]: I0227 16:43:53.062498 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pm95c" event={"ID":"c357e0a2-9571-4e9e-b48c-0f92e2c99afc","Type":"ContainerStarted","Data":"2b112b351ca9b51bbab80bb80a2b1bd6887725e81a1392ec79e65bdd4d6dbc0b"} Feb 27 16:43:53 crc kubenswrapper[4814]: I0227 16:43:53.064057 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-88h8r" podStartSLOduration=2.0640414910000002 podStartE2EDuration="2.064041491s" podCreationTimestamp="2026-02-27 16:43:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:43:53.061945274 +0000 UTC m=+1245.514570104" watchObservedRunningTime="2026-02-27 16:43:53.064041491 +0000 UTC m=+1245.516666321" Feb 27 16:43:53 crc kubenswrapper[4814]: I0227 16:43:53.066392 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01245f0c-38fe-4f9a-93e5-c5c5d5f32259","Type":"ContainerStarted","Data":"c52d992512b39f9030e3eb2b20244de9ced5f2493ef47f17365e81f5d87b6cf2"} Feb 27 16:43:53 crc kubenswrapper[4814]: I0227 16:43:53.096334 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5","Type":"ContainerStarted","Data":"82c9148f20c6fdb72bfaae8d94beaf97dd4c55f9689a736c2fcd251db6504b97"} Feb 27 16:43:53 crc kubenswrapper[4814]: I0227 16:43:53.103352 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-pm95c" podStartSLOduration=3.76669702 podStartE2EDuration="33.103336103s" podCreationTimestamp="2026-02-27 16:43:20 +0000 UTC" firstStartedPulling="2026-02-27 16:43:22.139528703 +0000 UTC m=+1214.592153533" lastFinishedPulling="2026-02-27 16:43:51.476167756 +0000 UTC m=+1243.928792616" observedRunningTime="2026-02-27 16:43:53.102698243 +0000 UTC m=+1245.555323073" watchObservedRunningTime="2026-02-27 16:43:53.103336103 +0000 UTC m=+1245.555960933" Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.125922 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65784f76f6-f2pcp" event={"ID":"8af960c4-8a04-42d8-83bf-9d03c23ad333","Type":"ContainerStarted","Data":"5df57cefa5e1c55606cf37ba1e71a0be1d58c6df74a16783ee38f0e64999beda"} Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.137655 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5","Type":"ContainerStarted","Data":"0c9eb449398eb3654fc7114dbf4cf0886a0c7ec6574f29d26e066a62c7c64417"} Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.144736 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-845d8f46cb-qhstj" event={"ID":"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1","Type":"ContainerStarted","Data":"5c624a9feb184d9603fa9c3597fae610b34d3632ba8c53b56374223a4f027fd2"} Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.144784 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-845d8f46cb-qhstj" event={"ID":"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1","Type":"ContainerStarted","Data":"19807c67c77d45f3562d439c93d9fa848a39b696dec9481c4ec2a11a35968141"} Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.154022 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-65784f76f6-f2pcp" podStartSLOduration=24.659072691 podStartE2EDuration="25.154000041s" podCreationTimestamp="2026-02-27 16:43:29 +0000 UTC" firstStartedPulling="2026-02-27 16:43:52.063211505 +0000 UTC m=+1244.515836335" lastFinishedPulling="2026-02-27 16:43:52.558138855 +0000 UTC m=+1245.010763685" observedRunningTime="2026-02-27 16:43:54.147226479 +0000 UTC m=+1246.599851309" watchObservedRunningTime="2026-02-27 16:43:54.154000041 +0000 UTC m=+1246.606624871" Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.158329 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31","Type":"ContainerStarted","Data":"04714ba3748af9f5fe9f795ce59413df20029c8ea6f5ebf98fe3748e51f45baf"} Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.158379 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31","Type":"ContainerStarted","Data":"6f9d88326b4353e59265198023144b951d9a4aad7e4934d48508f34f04a52792"} Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.184239 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-845d8f46cb-qhstj" podStartSLOduration=24.391931274 podStartE2EDuration="25.184211589s" podCreationTimestamp="2026-02-27 16:43:29 +0000 UTC" firstStartedPulling="2026-02-27 16:43:51.999524078 +0000 UTC m=+1244.452148898" lastFinishedPulling="2026-02-27 16:43:52.791804393 +0000 UTC m=+1245.244429213" observedRunningTime="2026-02-27 16:43:54.167698011 +0000 UTC m=+1246.620322831" watchObservedRunningTime="2026-02-27 16:43:54.184211589 +0000 UTC m=+1246.636836419" Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.192453 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=14.192427477 podStartE2EDuration="14.192427477s" podCreationTimestamp="2026-02-27 16:43:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:43:54.187079229 +0000 UTC m=+1246.639704069" watchObservedRunningTime="2026-02-27 16:43:54.192427477 +0000 UTC m=+1246.645052307" Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.702811 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-th62b" Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.839012 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5e413768-6d50-4725-a809-3658e913cfed-config\") pod \"5e413768-6d50-4725-a809-3658e913cfed\" (UID: \"5e413768-6d50-4725-a809-3658e913cfed\") " Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.839189 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e413768-6d50-4725-a809-3658e913cfed-combined-ca-bundle\") pod \"5e413768-6d50-4725-a809-3658e913cfed\" (UID: \"5e413768-6d50-4725-a809-3658e913cfed\") " Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.839239 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6xn5\" (UniqueName: \"kubernetes.io/projected/5e413768-6d50-4725-a809-3658e913cfed-kube-api-access-m6xn5\") pod \"5e413768-6d50-4725-a809-3658e913cfed\" (UID: \"5e413768-6d50-4725-a809-3658e913cfed\") " Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.863482 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e413768-6d50-4725-a809-3658e913cfed-kube-api-access-m6xn5" (OuterVolumeSpecName: "kube-api-access-m6xn5") pod "5e413768-6d50-4725-a809-3658e913cfed" (UID: "5e413768-6d50-4725-a809-3658e913cfed"). InnerVolumeSpecName "kube-api-access-m6xn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.873197 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e413768-6d50-4725-a809-3658e913cfed-config" (OuterVolumeSpecName: "config") pod "5e413768-6d50-4725-a809-3658e913cfed" (UID: "5e413768-6d50-4725-a809-3658e913cfed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.873363 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e413768-6d50-4725-a809-3658e913cfed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e413768-6d50-4725-a809-3658e913cfed" (UID: "5e413768-6d50-4725-a809-3658e913cfed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.941526 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5e413768-6d50-4725-a809-3658e913cfed-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.941583 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e413768-6d50-4725-a809-3658e913cfed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:54 crc kubenswrapper[4814]: I0227 16:43:54.941599 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6xn5\" (UniqueName: \"kubernetes.io/projected/5e413768-6d50-4725-a809-3658e913cfed-kube-api-access-m6xn5\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.173913 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5","Type":"ContainerStarted","Data":"d407574566976064af8e3d96a78617ed018b92a07e84a6d1ddc32338af3c2c3e"} Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.176882 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-th62b" event={"ID":"5e413768-6d50-4725-a809-3658e913cfed","Type":"ContainerDied","Data":"426d5abb10ada29b603a0a679e0c83ebbd8aa1defd09e7c392d89480b81a779e"} Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.176927 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="426d5abb10ada29b603a0a679e0c83ebbd8aa1defd09e7c392d89480b81a779e" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.177331 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-th62b" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.208030 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.208006805 podStartE2EDuration="4.208006805s" podCreationTimestamp="2026-02-27 16:43:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:43:55.205458635 +0000 UTC m=+1247.658083465" watchObservedRunningTime="2026-02-27 16:43:55.208006805 +0000 UTC m=+1247.660631645" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.342790 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-hx978"] Feb 27 16:43:55 crc kubenswrapper[4814]: E0227 16:43:55.343294 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e413768-6d50-4725-a809-3658e913cfed" containerName="neutron-db-sync" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.343319 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e413768-6d50-4725-a809-3658e913cfed" containerName="neutron-db-sync" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.343520 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e413768-6d50-4725-a809-3658e913cfed" containerName="neutron-db-sync" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.345971 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.353466 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-hx978"] Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.430380 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5f986b6894-p4tf6"] Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.431728 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.438066 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.438466 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-lhq6v" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.438782 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.439856 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.456261 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-config\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.456301 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-ovndb-tls-certs\") pod \"neutron-5f986b6894-p4tf6\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.456326 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-httpd-config\") pod \"neutron-5f986b6894-p4tf6\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.456395 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-config\") pod \"neutron-5f986b6894-p4tf6\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.456416 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj5x6\" (UniqueName: \"kubernetes.io/projected/883b18b4-e86b-4077-8865-ff24ae4838e1-kube-api-access-tj5x6\") pod \"neutron-5f986b6894-p4tf6\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.456444 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9prqt\" (UniqueName: \"kubernetes.io/projected/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-kube-api-access-9prqt\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.456459 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.456478 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.456499 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.456514 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-combined-ca-bundle\") pod \"neutron-5f986b6894-p4tf6\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.456556 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.461820 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5f986b6894-p4tf6"] Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.557588 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.557660 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-config\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.557678 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-ovndb-tls-certs\") pod \"neutron-5f986b6894-p4tf6\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.557698 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-httpd-config\") pod \"neutron-5f986b6894-p4tf6\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.557801 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-config\") pod \"neutron-5f986b6894-p4tf6\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.557830 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj5x6\" (UniqueName: \"kubernetes.io/projected/883b18b4-e86b-4077-8865-ff24ae4838e1-kube-api-access-tj5x6\") pod \"neutron-5f986b6894-p4tf6\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.557855 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9prqt\" (UniqueName: \"kubernetes.io/projected/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-kube-api-access-9prqt\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.557878 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.557896 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.557913 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.557932 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-combined-ca-bundle\") pod \"neutron-5f986b6894-p4tf6\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.563429 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.563474 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.564560 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-config\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.566796 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-httpd-config\") pod \"neutron-5f986b6894-p4tf6\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.574040 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-combined-ca-bundle\") pod \"neutron-5f986b6894-p4tf6\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.581908 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.582160 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-ovndb-tls-certs\") pod \"neutron-5f986b6894-p4tf6\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.583808 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.589081 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9prqt\" (UniqueName: \"kubernetes.io/projected/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-kube-api-access-9prqt\") pod \"dnsmasq-dns-84b966f6c9-hx978\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.592561 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-config\") pod \"neutron-5f986b6894-p4tf6\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.603947 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj5x6\" (UniqueName: \"kubernetes.io/projected/883b18b4-e86b-4077-8865-ff24ae4838e1-kube-api-access-tj5x6\") pod \"neutron-5f986b6894-p4tf6\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.680662 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:43:55 crc kubenswrapper[4814]: I0227 16:43:55.780072 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:43:56 crc kubenswrapper[4814]: I0227 16:43:56.197493 4814 generic.go:334] "Generic (PLEG): container finished" podID="c357e0a2-9571-4e9e-b48c-0f92e2c99afc" containerID="2b112b351ca9b51bbab80bb80a2b1bd6887725e81a1392ec79e65bdd4d6dbc0b" exitCode=0 Feb 27 16:43:56 crc kubenswrapper[4814]: I0227 16:43:56.197565 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pm95c" event={"ID":"c357e0a2-9571-4e9e-b48c-0f92e2c99afc","Type":"ContainerDied","Data":"2b112b351ca9b51bbab80bb80a2b1bd6887725e81a1392ec79e65bdd4d6dbc0b"} Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.206874 4814 generic.go:334] "Generic (PLEG): container finished" podID="6f56fd06-b678-48f2-9f0d-0234a53b528b" containerID="67fab44cb9ddab92dbc952ba5825863ced0e115a1f4e9dac622240cd0f1e0176" exitCode=0 Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.207422 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-88h8r" event={"ID":"6f56fd06-b678-48f2-9f0d-0234a53b528b","Type":"ContainerDied","Data":"67fab44cb9ddab92dbc952ba5825863ced0e115a1f4e9dac622240cd0f1e0176"} Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.583557 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pm95c" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.660550 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-85f77557cc-k22bs"] Feb 27 16:43:57 crc kubenswrapper[4814]: E0227 16:43:57.662378 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c357e0a2-9571-4e9e-b48c-0f92e2c99afc" containerName="barbican-db-sync" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.662406 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="c357e0a2-9571-4e9e-b48c-0f92e2c99afc" containerName="barbican-db-sync" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.662631 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="c357e0a2-9571-4e9e-b48c-0f92e2c99afc" containerName="barbican-db-sync" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.664048 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.667906 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.668082 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.674503 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-85f77557cc-k22bs"] Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.708253 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-db-sync-config-data\") pod \"c357e0a2-9571-4e9e-b48c-0f92e2c99afc\" (UID: \"c357e0a2-9571-4e9e-b48c-0f92e2c99afc\") " Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.708417 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-combined-ca-bundle\") pod \"c357e0a2-9571-4e9e-b48c-0f92e2c99afc\" (UID: \"c357e0a2-9571-4e9e-b48c-0f92e2c99afc\") " Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.708523 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsd8h\" (UniqueName: \"kubernetes.io/projected/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-kube-api-access-qsd8h\") pod \"c357e0a2-9571-4e9e-b48c-0f92e2c99afc\" (UID: \"c357e0a2-9571-4e9e-b48c-0f92e2c99afc\") " Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.709380 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-combined-ca-bundle\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.709486 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs5nw\" (UniqueName: \"kubernetes.io/projected/131dd92c-e25a-4385-9380-84e286f09a43-kube-api-access-xs5nw\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.709530 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-httpd-config\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.709575 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-config\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.709605 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-internal-tls-certs\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.709666 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-ovndb-tls-certs\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.709736 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-public-tls-certs\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.727629 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-kube-api-access-qsd8h" (OuterVolumeSpecName: "kube-api-access-qsd8h") pod "c357e0a2-9571-4e9e-b48c-0f92e2c99afc" (UID: "c357e0a2-9571-4e9e-b48c-0f92e2c99afc"). InnerVolumeSpecName "kube-api-access-qsd8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.727700 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c357e0a2-9571-4e9e-b48c-0f92e2c99afc" (UID: "c357e0a2-9571-4e9e-b48c-0f92e2c99afc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.762090 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c357e0a2-9571-4e9e-b48c-0f92e2c99afc" (UID: "c357e0a2-9571-4e9e-b48c-0f92e2c99afc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.810979 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-public-tls-certs\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.811066 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-combined-ca-bundle\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.811094 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs5nw\" (UniqueName: \"kubernetes.io/projected/131dd92c-e25a-4385-9380-84e286f09a43-kube-api-access-xs5nw\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.811111 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-httpd-config\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.811139 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-config\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.811157 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-internal-tls-certs\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.811197 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-ovndb-tls-certs\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.811257 4814 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.811268 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.811277 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsd8h\" (UniqueName: \"kubernetes.io/projected/c357e0a2-9571-4e9e-b48c-0f92e2c99afc-kube-api-access-qsd8h\") on node \"crc\" DevicePath \"\"" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.814987 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-ovndb-tls-certs\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.817467 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-httpd-config\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.819090 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-combined-ca-bundle\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.819971 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-public-tls-certs\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.830997 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-internal-tls-certs\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.831581 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs5nw\" (UniqueName: \"kubernetes.io/projected/131dd92c-e25a-4385-9380-84e286f09a43-kube-api-access-xs5nw\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.832689 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-config\") pod \"neutron-85f77557cc-k22bs\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:57 crc kubenswrapper[4814]: I0227 16:43:57.980391 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.215497 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pm95c" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.220469 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pm95c" event={"ID":"c357e0a2-9571-4e9e-b48c-0f92e2c99afc","Type":"ContainerDied","Data":"47c8485ae374b4e18fc8661595348d4ace5b324d6a797c81472eaf5e774aedc7"} Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.220498 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47c8485ae374b4e18fc8661595348d4ace5b324d6a797c81472eaf5e774aedc7" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.579435 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-84b49964c9-d96qj"] Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.580978 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.595236 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.595407 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.595546 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-ll2qr" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.596522 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-66c4c86998-k7mxw"] Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.597779 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.603974 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.617108 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-84b49964c9-d96qj"] Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.626299 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-66c4c86998-k7mxw"] Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.631929 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-config-data-custom\") pod \"barbican-worker-84b49964c9-d96qj\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.632392 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-config-data\") pod \"barbican-keystone-listener-66c4c86998-k7mxw\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.632526 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-logs\") pod \"barbican-worker-84b49964c9-d96qj\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.632601 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmsrk\" (UniqueName: \"kubernetes.io/projected/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-kube-api-access-mmsrk\") pod \"barbican-keystone-listener-66c4c86998-k7mxw\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.632700 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2f8l\" (UniqueName: \"kubernetes.io/projected/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-kube-api-access-q2f8l\") pod \"barbican-worker-84b49964c9-d96qj\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.632801 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-combined-ca-bundle\") pod \"barbican-keystone-listener-66c4c86998-k7mxw\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.632918 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-config-data\") pod \"barbican-worker-84b49964c9-d96qj\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.633007 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-config-data-custom\") pod \"barbican-keystone-listener-66c4c86998-k7mxw\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.633170 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-logs\") pod \"barbican-keystone-listener-66c4c86998-k7mxw\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.633247 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-combined-ca-bundle\") pod \"barbican-worker-84b49964c9-d96qj\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.734057 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-logs\") pod \"barbican-worker-84b49964c9-d96qj\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.734101 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmsrk\" (UniqueName: \"kubernetes.io/projected/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-kube-api-access-mmsrk\") pod \"barbican-keystone-listener-66c4c86998-k7mxw\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.734140 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2f8l\" (UniqueName: \"kubernetes.io/projected/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-kube-api-access-q2f8l\") pod \"barbican-worker-84b49964c9-d96qj\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.734183 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-combined-ca-bundle\") pod \"barbican-keystone-listener-66c4c86998-k7mxw\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.734211 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-config-data\") pod \"barbican-worker-84b49964c9-d96qj\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.734228 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-config-data-custom\") pod \"barbican-keystone-listener-66c4c86998-k7mxw\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.734245 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-logs\") pod \"barbican-keystone-listener-66c4c86998-k7mxw\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.734263 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-combined-ca-bundle\") pod \"barbican-worker-84b49964c9-d96qj\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.734308 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-config-data-custom\") pod \"barbican-worker-84b49964c9-d96qj\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.734335 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-config-data\") pod \"barbican-keystone-listener-66c4c86998-k7mxw\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.735270 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-logs\") pod \"barbican-worker-84b49964c9-d96qj\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.735655 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-logs\") pod \"barbican-keystone-listener-66c4c86998-k7mxw\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.741067 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-combined-ca-bundle\") pod \"barbican-worker-84b49964c9-d96qj\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.746796 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-config-data-custom\") pod \"barbican-worker-84b49964c9-d96qj\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.748164 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-config-data\") pod \"barbican-keystone-listener-66c4c86998-k7mxw\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.773188 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-config-data-custom\") pod \"barbican-keystone-listener-66c4c86998-k7mxw\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.773787 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-combined-ca-bundle\") pod \"barbican-keystone-listener-66c4c86998-k7mxw\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.776540 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-config-data\") pod \"barbican-worker-84b49964c9-d96qj\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.784520 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2f8l\" (UniqueName: \"kubernetes.io/projected/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-kube-api-access-q2f8l\") pod \"barbican-worker-84b49964c9-d96qj\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.796099 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmsrk\" (UniqueName: \"kubernetes.io/projected/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-kube-api-access-mmsrk\") pod \"barbican-keystone-listener-66c4c86998-k7mxw\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.801324 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-hx978"] Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.858146 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6f4999f464-wcs2l"] Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.868356 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.871674 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.883882 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6f4999f464-wcs2l"] Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.895162 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-2s2xt"] Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.896581 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.919380 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-2s2xt"] Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.928602 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.946551 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.947461 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-config-data\") pod \"barbican-api-6f4999f464-wcs2l\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.947517 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.947542 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-config-data-custom\") pod \"barbican-api-6f4999f464-wcs2l\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.947576 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-config\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.947606 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0b8557d-2397-4c7d-b6ab-6044db5f0129-logs\") pod \"barbican-api-6f4999f464-wcs2l\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.947672 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccrlt\" (UniqueName: \"kubernetes.io/projected/097c7fc7-f42b-494c-8d25-f7e76a8350b5-kube-api-access-ccrlt\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.947703 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-combined-ca-bundle\") pod \"barbican-api-6f4999f464-wcs2l\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.947728 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.947747 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv95z\" (UniqueName: \"kubernetes.io/projected/c0b8557d-2397-4c7d-b6ab-6044db5f0129-kube-api-access-jv95z\") pod \"barbican-api-6f4999f464-wcs2l\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.947771 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:58 crc kubenswrapper[4814]: I0227 16:43:58.947791 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.049190 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-config-data\") pod \"barbican-api-6f4999f464-wcs2l\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.049240 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.049257 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-config-data-custom\") pod \"barbican-api-6f4999f464-wcs2l\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.049308 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-config\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.049333 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0b8557d-2397-4c7d-b6ab-6044db5f0129-logs\") pod \"barbican-api-6f4999f464-wcs2l\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.049365 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccrlt\" (UniqueName: \"kubernetes.io/projected/097c7fc7-f42b-494c-8d25-f7e76a8350b5-kube-api-access-ccrlt\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.049398 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-combined-ca-bundle\") pod \"barbican-api-6f4999f464-wcs2l\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.049424 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.049443 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv95z\" (UniqueName: \"kubernetes.io/projected/c0b8557d-2397-4c7d-b6ab-6044db5f0129-kube-api-access-jv95z\") pod \"barbican-api-6f4999f464-wcs2l\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.049464 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.049487 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.054342 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.066705 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.066738 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-config\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.067319 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-combined-ca-bundle\") pod \"barbican-api-6f4999f464-wcs2l\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.067463 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0b8557d-2397-4c7d-b6ab-6044db5f0129-logs\") pod \"barbican-api-6f4999f464-wcs2l\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.067678 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.068422 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.069077 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-config-data-custom\") pod \"barbican-api-6f4999f464-wcs2l\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.073111 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv95z\" (UniqueName: \"kubernetes.io/projected/c0b8557d-2397-4c7d-b6ab-6044db5f0129-kube-api-access-jv95z\") pod \"barbican-api-6f4999f464-wcs2l\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.074266 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-config-data\") pod \"barbican-api-6f4999f464-wcs2l\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.089065 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccrlt\" (UniqueName: \"kubernetes.io/projected/097c7fc7-f42b-494c-8d25-f7e76a8350b5-kube-api-access-ccrlt\") pod \"dnsmasq-dns-75c8ddd69c-2s2xt\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.192442 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.221777 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.670636 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.670675 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.763831 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:43:59 crc kubenswrapper[4814]: I0227 16:43:59.763883 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:44:00 crc kubenswrapper[4814]: I0227 16:44:00.125889 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536844-4sfb5"] Feb 27 16:44:00 crc kubenswrapper[4814]: I0227 16:44:00.126950 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536844-4sfb5" Feb 27 16:44:00 crc kubenswrapper[4814]: I0227 16:44:00.135080 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536844-4sfb5"] Feb 27 16:44:00 crc kubenswrapper[4814]: I0227 16:44:00.136207 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 16:44:00 crc kubenswrapper[4814]: I0227 16:44:00.136312 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:44:00 crc kubenswrapper[4814]: I0227 16:44:00.136312 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:44:00 crc kubenswrapper[4814]: I0227 16:44:00.166816 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqbc6\" (UniqueName: \"kubernetes.io/projected/68c41339-6536-4998-901e-179aee23ac8c-kube-api-access-kqbc6\") pod \"auto-csr-approver-29536844-4sfb5\" (UID: \"68c41339-6536-4998-901e-179aee23ac8c\") " pod="openshift-infra/auto-csr-approver-29536844-4sfb5" Feb 27 16:44:00 crc kubenswrapper[4814]: I0227 16:44:00.268485 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqbc6\" (UniqueName: \"kubernetes.io/projected/68c41339-6536-4998-901e-179aee23ac8c-kube-api-access-kqbc6\") pod \"auto-csr-approver-29536844-4sfb5\" (UID: \"68c41339-6536-4998-901e-179aee23ac8c\") " pod="openshift-infra/auto-csr-approver-29536844-4sfb5" Feb 27 16:44:00 crc kubenswrapper[4814]: I0227 16:44:00.288068 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqbc6\" (UniqueName: \"kubernetes.io/projected/68c41339-6536-4998-901e-179aee23ac8c-kube-api-access-kqbc6\") pod \"auto-csr-approver-29536844-4sfb5\" (UID: \"68c41339-6536-4998-901e-179aee23ac8c\") " pod="openshift-infra/auto-csr-approver-29536844-4sfb5" Feb 27 16:44:00 crc kubenswrapper[4814]: I0227 16:44:00.445178 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536844-4sfb5" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.490990 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.491330 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.559942 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.572908 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.575716 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.575752 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.599846 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.648733 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.681614 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.697702 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-fernet-keys\") pod \"6f56fd06-b678-48f2-9f0d-0234a53b528b\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.697800 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-scripts\") pod \"6f56fd06-b678-48f2-9f0d-0234a53b528b\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.698981 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-combined-ca-bundle\") pod \"6f56fd06-b678-48f2-9f0d-0234a53b528b\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.699086 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-config-data\") pod \"6f56fd06-b678-48f2-9f0d-0234a53b528b\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.699125 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-credential-keys\") pod \"6f56fd06-b678-48f2-9f0d-0234a53b528b\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.699159 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v94rk\" (UniqueName: \"kubernetes.io/projected/6f56fd06-b678-48f2-9f0d-0234a53b528b-kube-api-access-v94rk\") pod \"6f56fd06-b678-48f2-9f0d-0234a53b528b\" (UID: \"6f56fd06-b678-48f2-9f0d-0234a53b528b\") " Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.710859 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6f56fd06-b678-48f2-9f0d-0234a53b528b" (UID: "6f56fd06-b678-48f2-9f0d-0234a53b528b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.725418 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6f56fd06-b678-48f2-9f0d-0234a53b528b" (UID: "6f56fd06-b678-48f2-9f0d-0234a53b528b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.730702 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-scripts" (OuterVolumeSpecName: "scripts") pod "6f56fd06-b678-48f2-9f0d-0234a53b528b" (UID: "6f56fd06-b678-48f2-9f0d-0234a53b528b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.731780 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f56fd06-b678-48f2-9f0d-0234a53b528b-kube-api-access-v94rk" (OuterVolumeSpecName: "kube-api-access-v94rk") pod "6f56fd06-b678-48f2-9f0d-0234a53b528b" (UID: "6f56fd06-b678-48f2-9f0d-0234a53b528b"). InnerVolumeSpecName "kube-api-access-v94rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.802575 4814 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.802602 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v94rk\" (UniqueName: \"kubernetes.io/projected/6f56fd06-b678-48f2-9f0d-0234a53b528b-kube-api-access-v94rk\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.802613 4814 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.802621 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.813056 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-config-data" (OuterVolumeSpecName: "config-data") pod "6f56fd06-b678-48f2-9f0d-0234a53b528b" (UID: "6f56fd06-b678-48f2-9f0d-0234a53b528b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.837749 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f56fd06-b678-48f2-9f0d-0234a53b528b" (UID: "6f56fd06-b678-48f2-9f0d-0234a53b528b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.904348 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:01 crc kubenswrapper[4814]: I0227 16:44:01.904374 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f56fd06-b678-48f2-9f0d-0234a53b528b-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.092950 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-855bdc9fdb-m79h8"] Feb 27 16:44:02 crc kubenswrapper[4814]: E0227 16:44:02.094842 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f56fd06-b678-48f2-9f0d-0234a53b528b" containerName="keystone-bootstrap" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.094861 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f56fd06-b678-48f2-9f0d-0234a53b528b" containerName="keystone-bootstrap" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.095056 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f56fd06-b678-48f2-9f0d-0234a53b528b" containerName="keystone-bootstrap" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.095981 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.104150 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.104384 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.119213 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-84b49964c9-d96qj"] Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.178187 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-855bdc9fdb-m79h8"] Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.211248 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-combined-ca-bundle\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.211317 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-config-data\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.211342 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-public-tls-certs\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.211402 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-internal-tls-certs\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.211677 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98s7g\" (UniqueName: \"kubernetes.io/projected/3584eaa9-9218-4dca-848c-1781f316e4a2-kube-api-access-98s7g\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.211722 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-config-data-custom\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.211893 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3584eaa9-9218-4dca-848c-1781f316e4a2-logs\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.262564 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84b49964c9-d96qj" event={"ID":"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e","Type":"ContainerStarted","Data":"459674b1b87d9fd9bbd1c8e019783ce1064ec0a6b2352851c8b1e64c46e23614"} Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.267345 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-88h8r" event={"ID":"6f56fd06-b678-48f2-9f0d-0234a53b528b","Type":"ContainerDied","Data":"1a23ac6768bf39d8b065376f92384a96dac2703a6e8825bb53b875bcc595900d"} Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.267402 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a23ac6768bf39d8b065376f92384a96dac2703a6e8825bb53b875bcc595900d" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.267489 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-88h8r" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.269674 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.269709 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.269719 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.269816 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.313604 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3584eaa9-9218-4dca-848c-1781f316e4a2-logs\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.313993 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-combined-ca-bundle\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.314042 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-config-data\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.314058 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-public-tls-certs\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.314084 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-internal-tls-certs\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.314171 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98s7g\" (UniqueName: \"kubernetes.io/projected/3584eaa9-9218-4dca-848c-1781f316e4a2-kube-api-access-98s7g\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.314195 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-config-data-custom\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.319969 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-internal-tls-certs\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.314164 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3584eaa9-9218-4dca-848c-1781f316e4a2-logs\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.324960 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-public-tls-certs\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.327890 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-config-data-custom\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.328572 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-combined-ca-bundle\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.341352 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-config-data\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.350882 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98s7g\" (UniqueName: \"kubernetes.io/projected/3584eaa9-9218-4dca-848c-1781f316e4a2-kube-api-access-98s7g\") pod \"barbican-api-855bdc9fdb-m79h8\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.476045 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-hx978"] Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.573281 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.626831 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5f986b6894-p4tf6"] Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.758890 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7877c8f6d6-q56qs"] Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.760001 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.763495 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.763535 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.763497 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.763684 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.763740 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-m5bsz" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.763891 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.783908 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7877c8f6d6-q56qs"] Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.893442 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-85f77557cc-k22bs"] Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.920009 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6f4999f464-wcs2l"] Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.941740 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-internal-tls-certs\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.941778 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-combined-ca-bundle\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.941813 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-fernet-keys\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.941842 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f68qm\" (UniqueName: \"kubernetes.io/projected/017f253d-4956-4f30-93f1-66ead39f9980-kube-api-access-f68qm\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.941862 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-scripts\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.941890 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-credential-keys\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.942958 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-config-data\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.942997 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-public-tls-certs\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.945829 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-66c4c86998-k7mxw"] Feb 27 16:44:02 crc kubenswrapper[4814]: I0227 16:44:02.994149 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-2s2xt"] Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.006069 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536844-4sfb5"] Feb 27 16:44:03 crc kubenswrapper[4814]: W0227 16:44:03.019901 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod097c7fc7_f42b_494c_8d25_f7e76a8350b5.slice/crio-21ede465b3046ed6554972811baa18a99ff32a95bc2656e6a3a249ab80a89970 WatchSource:0}: Error finding container 21ede465b3046ed6554972811baa18a99ff32a95bc2656e6a3a249ab80a89970: Status 404 returned error can't find the container with id 21ede465b3046ed6554972811baa18a99ff32a95bc2656e6a3a249ab80a89970 Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.044030 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-internal-tls-certs\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.044440 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-combined-ca-bundle\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.044481 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-fernet-keys\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.044510 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f68qm\" (UniqueName: \"kubernetes.io/projected/017f253d-4956-4f30-93f1-66ead39f9980-kube-api-access-f68qm\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.044535 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-scripts\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.044561 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-credential-keys\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.044580 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-public-tls-certs\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.044595 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-config-data\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.049689 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-scripts\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.049745 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-internal-tls-certs\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.049861 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-combined-ca-bundle\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.049872 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-public-tls-certs\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.051815 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-config-data\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.052905 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-credential-keys\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.053021 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/017f253d-4956-4f30-93f1-66ead39f9980-fernet-keys\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.075095 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f68qm\" (UniqueName: \"kubernetes.io/projected/017f253d-4956-4f30-93f1-66ead39f9980-kube-api-access-f68qm\") pod \"keystone-7877c8f6d6-q56qs\" (UID: \"017f253d-4956-4f30-93f1-66ead39f9980\") " pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.117893 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.316656 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01245f0c-38fe-4f9a-93e5-c5c5d5f32259","Type":"ContainerStarted","Data":"fe992a34a5bd33198fa76c1d351deb8c6a8cb1d10babbc60cf65c741734db762"} Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.322539 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" event={"ID":"097c7fc7-f42b-494c-8d25-f7e76a8350b5","Type":"ContainerStarted","Data":"21ede465b3046ed6554972811baa18a99ff32a95bc2656e6a3a249ab80a89970"} Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.369027 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" event={"ID":"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b","Type":"ContainerStarted","Data":"0dbb3beb7b5d2c86451925d2da96ad6882e8810a5a85fcd89a359e636134459e"} Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.374764 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-855bdc9fdb-m79h8"] Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.376548 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f986b6894-p4tf6" event={"ID":"883b18b4-e86b-4077-8865-ff24ae4838e1","Type":"ContainerStarted","Data":"cfa61f32aebd404721137338d8249be5279fd83f1dab45436e53bbca1961e27c"} Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.376603 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f986b6894-p4tf6" event={"ID":"883b18b4-e86b-4077-8865-ff24ae4838e1","Type":"ContainerStarted","Data":"c05911f14d7f1b2fa25760f3ec14493754772f5170d78ea23abdd7c1a6e9e453"} Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.396999 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f4999f464-wcs2l" event={"ID":"c0b8557d-2397-4c7d-b6ab-6044db5f0129","Type":"ContainerStarted","Data":"8a61d2ec28d34e888e8723543aaf09c8b1750abf0dcc6b37ec5dd3c7870e764c"} Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.423896 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pq8r4" event={"ID":"6f5812c2-9952-4125-b055-dd2888fad3f5","Type":"ContainerStarted","Data":"f8e24f09444ee738f777761fa58406e7797de5453d3ece79e9a37b9268433aec"} Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.428469 4814 generic.go:334] "Generic (PLEG): container finished" podID="3efc1a92-93e1-45a5-86c6-263ad2d4c36b" containerID="6803bd37506d2301fc079cdf7a7032dfdfef505e0b1012a99096da020ba514cf" exitCode=0 Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.428527 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-hx978" event={"ID":"3efc1a92-93e1-45a5-86c6-263ad2d4c36b","Type":"ContainerDied","Data":"6803bd37506d2301fc079cdf7a7032dfdfef505e0b1012a99096da020ba514cf"} Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.428552 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-hx978" event={"ID":"3efc1a92-93e1-45a5-86c6-263ad2d4c36b","Type":"ContainerStarted","Data":"3b3de6af436cf391e032b4aff6fe2c4cd1f400b9369d6b4a15de75bb2ec24e50"} Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.430086 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536844-4sfb5" event={"ID":"68c41339-6536-4998-901e-179aee23ac8c","Type":"ContainerStarted","Data":"0b29d5a6c7ff0b2ca154c6aade40fa313328022e94c2593a052be3d8bc010bbf"} Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.431790 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85f77557cc-k22bs" event={"ID":"131dd92c-e25a-4385-9380-84e286f09a43","Type":"ContainerStarted","Data":"f5017af6bcc2517f9c5074da11f32164cbb6bd26a0dd518f692e885f3b43d209"} Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.440652 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-pq8r4" podStartSLOduration=3.683672007 podStartE2EDuration="43.440636306s" podCreationTimestamp="2026-02-27 16:43:20 +0000 UTC" firstStartedPulling="2026-02-27 16:43:22.157683052 +0000 UTC m=+1214.610307882" lastFinishedPulling="2026-02-27 16:44:01.914647351 +0000 UTC m=+1254.367272181" observedRunningTime="2026-02-27 16:44:03.439029106 +0000 UTC m=+1255.891653936" watchObservedRunningTime="2026-02-27 16:44:03.440636306 +0000 UTC m=+1255.893261136" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.580203 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5f986b6894-p4tf6"] Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.643323 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-69895c5d4f-4bpj5"] Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.644770 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.685632 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69895c5d4f-4bpj5"] Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.761285 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-public-tls-certs\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.761334 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-internal-tls-certs\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.761357 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-ovndb-tls-certs\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.761405 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-config\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.761422 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpfm6\" (UniqueName: \"kubernetes.io/projected/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-kube-api-access-rpfm6\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.761448 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-httpd-config\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.761467 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-combined-ca-bundle\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.867480 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-public-tls-certs\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.867525 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-internal-tls-certs\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.867543 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-ovndb-tls-certs\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.867594 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-config\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.867614 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpfm6\" (UniqueName: \"kubernetes.io/projected/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-kube-api-access-rpfm6\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.867644 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-httpd-config\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.867661 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-combined-ca-bundle\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.875005 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-config\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.876704 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-httpd-config\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.898511 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-combined-ca-bundle\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.899751 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-ovndb-tls-certs\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.900193 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-public-tls-certs\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.928877 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-internal-tls-certs\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:03 crc kubenswrapper[4814]: I0227 16:44:03.969143 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpfm6\" (UniqueName: \"kubernetes.io/projected/9e55471d-9b78-4e06-8195-c73f4ce0d1f6-kube-api-access-rpfm6\") pod \"neutron-69895c5d4f-4bpj5\" (UID: \"9e55471d-9b78-4e06-8195-c73f4ce0d1f6\") " pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.007042 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7877c8f6d6-q56qs"] Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.093402 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.105675 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.327741 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-ovsdbserver-sb\") pod \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.327829 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-dns-swift-storage-0\") pod \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.327931 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-ovsdbserver-nb\") pod \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.327963 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-dns-svc\") pod \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.327994 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9prqt\" (UniqueName: \"kubernetes.io/projected/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-kube-api-access-9prqt\") pod \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.328026 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-config\") pod \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\" (UID: \"3efc1a92-93e1-45a5-86c6-263ad2d4c36b\") " Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.341726 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-kube-api-access-9prqt" (OuterVolumeSpecName: "kube-api-access-9prqt") pod "3efc1a92-93e1-45a5-86c6-263ad2d4c36b" (UID: "3efc1a92-93e1-45a5-86c6-263ad2d4c36b"). InnerVolumeSpecName "kube-api-access-9prqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.403184 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3efc1a92-93e1-45a5-86c6-263ad2d4c36b" (UID: "3efc1a92-93e1-45a5-86c6-263ad2d4c36b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.416127 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3efc1a92-93e1-45a5-86c6-263ad2d4c36b" (UID: "3efc1a92-93e1-45a5-86c6-263ad2d4c36b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.416541 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3efc1a92-93e1-45a5-86c6-263ad2d4c36b" (UID: "3efc1a92-93e1-45a5-86c6-263ad2d4c36b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.429501 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9prqt\" (UniqueName: \"kubernetes.io/projected/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-kube-api-access-9prqt\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.429522 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.429531 4814 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.429539 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.441020 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3efc1a92-93e1-45a5-86c6-263ad2d4c36b" (UID: "3efc1a92-93e1-45a5-86c6-263ad2d4c36b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.445385 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-config" (OuterVolumeSpecName: "config") pod "3efc1a92-93e1-45a5-86c6-263ad2d4c36b" (UID: "3efc1a92-93e1-45a5-86c6-263ad2d4c36b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.478446 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7877c8f6d6-q56qs" event={"ID":"017f253d-4956-4f30-93f1-66ead39f9980","Type":"ContainerStarted","Data":"fb388b80e8a6f9a87a638cc14fae3bf3adfa9a47500f4b0933d21b092f7ff3b1"} Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.480544 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-855bdc9fdb-m79h8" event={"ID":"3584eaa9-9218-4dca-848c-1781f316e4a2","Type":"ContainerStarted","Data":"d2f2e337a2622ede196d96b51299a1cb5497818091db374683707fd677c69485"} Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.480564 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-855bdc9fdb-m79h8" event={"ID":"3584eaa9-9218-4dca-848c-1781f316e4a2","Type":"ContainerStarted","Data":"1b2c9e11000a8e15321ebb9234f502702be29180b7977e4a1c31bd2fc8e87b7e"} Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.481620 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-hx978" event={"ID":"3efc1a92-93e1-45a5-86c6-263ad2d4c36b","Type":"ContainerDied","Data":"3b3de6af436cf391e032b4aff6fe2c4cd1f400b9369d6b4a15de75bb2ec24e50"} Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.481645 4814 scope.go:117] "RemoveContainer" containerID="6803bd37506d2301fc079cdf7a7032dfdfef505e0b1012a99096da020ba514cf" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.481749 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.507130 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85f77557cc-k22bs" event={"ID":"131dd92c-e25a-4385-9380-84e286f09a43","Type":"ContainerStarted","Data":"bdaf42af64df4a5bbcec9295585e7515e63e2138955c345374e7af4e4023a5e8"} Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.507182 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85f77557cc-k22bs" event={"ID":"131dd92c-e25a-4385-9380-84e286f09a43","Type":"ContainerStarted","Data":"d13649093c51431f472803e631c21e26b1fdd0d8c979de1fc946481976409c01"} Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.507528 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.538632 4814 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.538688 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3efc1a92-93e1-45a5-86c6-263ad2d4c36b-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.610084 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f4999f464-wcs2l" event={"ID":"c0b8557d-2397-4c7d-b6ab-6044db5f0129","Type":"ContainerStarted","Data":"2421ada8cb323269d6db5f817fde7424f4db0ab7c760ab8bd82858da18be7108"} Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.610226 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f4999f464-wcs2l" event={"ID":"c0b8557d-2397-4c7d-b6ab-6044db5f0129","Type":"ContainerStarted","Data":"bf258e08576cf85838807e3706331bb269ce81b0024fdb537b74ec012819b4ba"} Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.610366 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.610387 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.682523 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f986b6894-p4tf6" event={"ID":"883b18b4-e86b-4077-8865-ff24ae4838e1","Type":"ContainerStarted","Data":"993646e01cc33127d18c148d44f285e775bc2a6765cd92bfeaf24782ce055cf5"} Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.698987 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-85f77557cc-k22bs" podStartSLOduration=7.698970117 podStartE2EDuration="7.698970117s" podCreationTimestamp="2026-02-27 16:43:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:04.536711638 +0000 UTC m=+1256.989336468" watchObservedRunningTime="2026-02-27 16:44:04.698970117 +0000 UTC m=+1257.151594947" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.699420 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5f986b6894-p4tf6" podUID="883b18b4-e86b-4077-8865-ff24ae4838e1" containerName="neutron-api" containerID="cri-o://cfa61f32aebd404721137338d8249be5279fd83f1dab45436e53bbca1961e27c" gracePeriod=30 Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.699763 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6f4999f464-wcs2l" podStartSLOduration=6.699756411 podStartE2EDuration="6.699756411s" podCreationTimestamp="2026-02-27 16:43:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:04.684328667 +0000 UTC m=+1257.136953497" watchObservedRunningTime="2026-02-27 16:44:04.699756411 +0000 UTC m=+1257.152381241" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.699869 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5f986b6894-p4tf6" podUID="883b18b4-e86b-4077-8865-ff24ae4838e1" containerName="neutron-httpd" containerID="cri-o://993646e01cc33127d18c148d44f285e775bc2a6765cd92bfeaf24782ce055cf5" gracePeriod=30 Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.723107 4814 generic.go:334] "Generic (PLEG): container finished" podID="097c7fc7-f42b-494c-8d25-f7e76a8350b5" containerID="3e12afe869bb6c7cd7c682a61ea3c6f27c1f7af9fce87a9707cffd7869f3b249" exitCode=0 Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.723206 4814 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.723216 4814 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.731600 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" event={"ID":"097c7fc7-f42b-494c-8d25-f7e76a8350b5","Type":"ContainerDied","Data":"3e12afe869bb6c7cd7c682a61ea3c6f27c1f7af9fce87a9707cffd7869f3b249"} Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.732066 4814 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.732079 4814 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.772552 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5f986b6894-p4tf6" podStartSLOduration=9.772531573 podStartE2EDuration="9.772531573s" podCreationTimestamp="2026-02-27 16:43:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:04.728025697 +0000 UTC m=+1257.180650527" watchObservedRunningTime="2026-02-27 16:44:04.772531573 +0000 UTC m=+1257.225156393" Feb 27 16:44:04 crc kubenswrapper[4814]: I0227 16:44:04.890905 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69895c5d4f-4bpj5"] Feb 27 16:44:04 crc kubenswrapper[4814]: W0227 16:44:04.979223 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e55471d_9b78_4e06_8195_c73f4ce0d1f6.slice/crio-4fee7f6b91c14ae142a3e44eb21120c6ec2a13526a7043204ac6b040e5ca3242 WatchSource:0}: Error finding container 4fee7f6b91c14ae142a3e44eb21120c6ec2a13526a7043204ac6b040e5ca3242: Status 404 returned error can't find the container with id 4fee7f6b91c14ae142a3e44eb21120c6ec2a13526a7043204ac6b040e5ca3242 Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.082038 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.095631 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.500585 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.777199 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-855bdc9fdb-m79h8" event={"ID":"3584eaa9-9218-4dca-848c-1781f316e4a2","Type":"ContainerStarted","Data":"f36a811f0c7b74a3b41ec489863d52d12d465987224d9968a5e7b15e60acf8ec"} Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.778619 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.778645 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.792702 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69895c5d4f-4bpj5" event={"ID":"9e55471d-9b78-4e06-8195-c73f4ce0d1f6","Type":"ContainerStarted","Data":"5db9acdbb67430dff1a8d46cc8bd722c32a93bbacbee1f81818ce5acf07a533b"} Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.792744 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69895c5d4f-4bpj5" event={"ID":"9e55471d-9b78-4e06-8195-c73f4ce0d1f6","Type":"ContainerStarted","Data":"4fee7f6b91c14ae142a3e44eb21120c6ec2a13526a7043204ac6b040e5ca3242"} Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.810864 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-855bdc9fdb-m79h8" podStartSLOduration=3.810842555 podStartE2EDuration="3.810842555s" podCreationTimestamp="2026-02-27 16:44:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:05.797861407 +0000 UTC m=+1258.250486227" watchObservedRunningTime="2026-02-27 16:44:05.810842555 +0000 UTC m=+1258.263467385" Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.811189 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536844-4sfb5" event={"ID":"68c41339-6536-4998-901e-179aee23ac8c","Type":"ContainerStarted","Data":"6eb8cd24d369d2d009c45dc31fe5e0ecd17dd117cae0420b6edc105714bbbd82"} Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.824921 4814 generic.go:334] "Generic (PLEG): container finished" podID="883b18b4-e86b-4077-8865-ff24ae4838e1" containerID="993646e01cc33127d18c148d44f285e775bc2a6765cd92bfeaf24782ce055cf5" exitCode=0 Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.824985 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f986b6894-p4tf6" event={"ID":"883b18b4-e86b-4077-8865-ff24ae4838e1","Type":"ContainerDied","Data":"993646e01cc33127d18c148d44f285e775bc2a6765cd92bfeaf24782ce055cf5"} Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.839056 4814 generic.go:334] "Generic (PLEG): container finished" podID="6f5812c2-9952-4125-b055-dd2888fad3f5" containerID="f8e24f09444ee738f777761fa58406e7797de5453d3ece79e9a37b9268433aec" exitCode=0 Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.839106 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pq8r4" event={"ID":"6f5812c2-9952-4125-b055-dd2888fad3f5","Type":"ContainerDied","Data":"f8e24f09444ee738f777761fa58406e7797de5453d3ece79e9a37b9268433aec"} Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.842534 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536844-4sfb5" podStartSLOduration=4.760881777 podStartE2EDuration="5.842513687s" podCreationTimestamp="2026-02-27 16:44:00 +0000 UTC" firstStartedPulling="2026-02-27 16:44:03.001914188 +0000 UTC m=+1255.454539018" lastFinishedPulling="2026-02-27 16:44:04.083546098 +0000 UTC m=+1256.536170928" observedRunningTime="2026-02-27 16:44:05.826300129 +0000 UTC m=+1258.278924959" watchObservedRunningTime="2026-02-27 16:44:05.842513687 +0000 UTC m=+1258.295138517" Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.865657 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" event={"ID":"097c7fc7-f42b-494c-8d25-f7e76a8350b5","Type":"ContainerStarted","Data":"aeee546fd66f4c8b06a06463a4c7549770e19ff37831945befcd27b375a8a0f8"} Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.866657 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.898531 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" podStartSLOduration=7.898514953 podStartE2EDuration="7.898514953s" podCreationTimestamp="2026-02-27 16:43:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:05.892074292 +0000 UTC m=+1258.344699122" watchObservedRunningTime="2026-02-27 16:44:05.898514953 +0000 UTC m=+1258.351139783" Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.900240 4814 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.900591 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7877c8f6d6-q56qs" event={"ID":"017f253d-4956-4f30-93f1-66ead39f9980","Type":"ContainerStarted","Data":"40a5f1b0141262361500ace91c41c8c9db064524d20ef9a7f9ffd30a58e2cd46"} Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.901631 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:05 crc kubenswrapper[4814]: I0227 16:44:05.967242 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7877c8f6d6-q56qs" podStartSLOduration=3.967223399 podStartE2EDuration="3.967223399s" podCreationTimestamp="2026-02-27 16:44:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:05.952542618 +0000 UTC m=+1258.405167448" watchObservedRunningTime="2026-02-27 16:44:05.967223399 +0000 UTC m=+1258.419848229" Feb 27 16:44:06 crc kubenswrapper[4814]: I0227 16:44:06.130175 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 27 16:44:06 crc kubenswrapper[4814]: I0227 16:44:06.953715 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69895c5d4f-4bpj5" event={"ID":"9e55471d-9b78-4e06-8195-c73f4ce0d1f6","Type":"ContainerStarted","Data":"41cfba3fc5559076e58168a7acd51580360e93d5cba64a9945d1c7918f5b86ca"} Feb 27 16:44:06 crc kubenswrapper[4814]: I0227 16:44:06.954473 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:06 crc kubenswrapper[4814]: I0227 16:44:06.968978 4814 generic.go:334] "Generic (PLEG): container finished" podID="68c41339-6536-4998-901e-179aee23ac8c" containerID="6eb8cd24d369d2d009c45dc31fe5e0ecd17dd117cae0420b6edc105714bbbd82" exitCode=0 Feb 27 16:44:06 crc kubenswrapper[4814]: I0227 16:44:06.969161 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536844-4sfb5" event={"ID":"68c41339-6536-4998-901e-179aee23ac8c","Type":"ContainerDied","Data":"6eb8cd24d369d2d009c45dc31fe5e0ecd17dd117cae0420b6edc105714bbbd82"} Feb 27 16:44:06 crc kubenswrapper[4814]: I0227 16:44:06.989369 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-69895c5d4f-4bpj5" podStartSLOduration=3.989349442 podStartE2EDuration="3.989349442s" podCreationTimestamp="2026-02-27 16:44:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:06.987651468 +0000 UTC m=+1259.440276298" watchObservedRunningTime="2026-02-27 16:44:06.989349442 +0000 UTC m=+1259.441974282" Feb 27 16:44:07 crc kubenswrapper[4814]: I0227 16:44:07.979818 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vqkl5" event={"ID":"7bcf9bc6-1fb4-47b4-b12d-017902e4992e","Type":"ContainerStarted","Data":"5c90611e97642d06afcbae7dc9e52e2df50a2e1faaa225ef508016d671dd756f"} Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.002386 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-vqkl5" podStartSLOduration=4.047836577 podStartE2EDuration="48.00237152s" podCreationTimestamp="2026-02-27 16:43:20 +0000 UTC" firstStartedPulling="2026-02-27 16:43:22.141100772 +0000 UTC m=+1214.593725602" lastFinishedPulling="2026-02-27 16:44:06.095635715 +0000 UTC m=+1258.548260545" observedRunningTime="2026-02-27 16:44:07.995795433 +0000 UTC m=+1260.448420263" watchObservedRunningTime="2026-02-27 16:44:08.00237152 +0000 UTC m=+1260.454996350" Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.460240 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pq8r4" Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.593040 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-config-data\") pod \"6f5812c2-9952-4125-b055-dd2888fad3f5\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.593470 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-combined-ca-bundle\") pod \"6f5812c2-9952-4125-b055-dd2888fad3f5\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.593631 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f5812c2-9952-4125-b055-dd2888fad3f5-logs\") pod \"6f5812c2-9952-4125-b055-dd2888fad3f5\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.593665 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-scripts\") pod \"6f5812c2-9952-4125-b055-dd2888fad3f5\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.593687 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k72fh\" (UniqueName: \"kubernetes.io/projected/6f5812c2-9952-4125-b055-dd2888fad3f5-kube-api-access-k72fh\") pod \"6f5812c2-9952-4125-b055-dd2888fad3f5\" (UID: \"6f5812c2-9952-4125-b055-dd2888fad3f5\") " Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.595475 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536844-4sfb5" Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.601322 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f5812c2-9952-4125-b055-dd2888fad3f5-logs" (OuterVolumeSpecName: "logs") pod "6f5812c2-9952-4125-b055-dd2888fad3f5" (UID: "6f5812c2-9952-4125-b055-dd2888fad3f5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.609704 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f5812c2-9952-4125-b055-dd2888fad3f5-kube-api-access-k72fh" (OuterVolumeSpecName: "kube-api-access-k72fh") pod "6f5812c2-9952-4125-b055-dd2888fad3f5" (UID: "6f5812c2-9952-4125-b055-dd2888fad3f5"). InnerVolumeSpecName "kube-api-access-k72fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.612408 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-scripts" (OuterVolumeSpecName: "scripts") pod "6f5812c2-9952-4125-b055-dd2888fad3f5" (UID: "6f5812c2-9952-4125-b055-dd2888fad3f5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.635486 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f5812c2-9952-4125-b055-dd2888fad3f5" (UID: "6f5812c2-9952-4125-b055-dd2888fad3f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.667052 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-config-data" (OuterVolumeSpecName: "config-data") pod "6f5812c2-9952-4125-b055-dd2888fad3f5" (UID: "6f5812c2-9952-4125-b055-dd2888fad3f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.698228 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqbc6\" (UniqueName: \"kubernetes.io/projected/68c41339-6536-4998-901e-179aee23ac8c-kube-api-access-kqbc6\") pod \"68c41339-6536-4998-901e-179aee23ac8c\" (UID: \"68c41339-6536-4998-901e-179aee23ac8c\") " Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.698734 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f5812c2-9952-4125-b055-dd2888fad3f5-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.698748 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.698756 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k72fh\" (UniqueName: \"kubernetes.io/projected/6f5812c2-9952-4125-b055-dd2888fad3f5-kube-api-access-k72fh\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.698766 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.698775 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f5812c2-9952-4125-b055-dd2888fad3f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.703734 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68c41339-6536-4998-901e-179aee23ac8c-kube-api-access-kqbc6" (OuterVolumeSpecName: "kube-api-access-kqbc6") pod "68c41339-6536-4998-901e-179aee23ac8c" (UID: "68c41339-6536-4998-901e-179aee23ac8c"). InnerVolumeSpecName "kube-api-access-kqbc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.801645 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqbc6\" (UniqueName: \"kubernetes.io/projected/68c41339-6536-4998-901e-179aee23ac8c-kube-api-access-kqbc6\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.931930 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536838-6bsv2"] Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.951275 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536838-6bsv2"] Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.995158 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536844-4sfb5" event={"ID":"68c41339-6536-4998-901e-179aee23ac8c","Type":"ContainerDied","Data":"0b29d5a6c7ff0b2ca154c6aade40fa313328022e94c2593a052be3d8bc010bbf"} Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.995202 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b29d5a6c7ff0b2ca154c6aade40fa313328022e94c2593a052be3d8bc010bbf" Feb 27 16:44:08 crc kubenswrapper[4814]: I0227 16:44:08.995291 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536844-4sfb5" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.000025 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" event={"ID":"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b","Type":"ContainerStarted","Data":"6a28d6cd48600190d077b9746978ba0bd442982fe85401efd2b8470fe849af70"} Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.002844 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-pq8r4" event={"ID":"6f5812c2-9952-4125-b055-dd2888fad3f5","Type":"ContainerDied","Data":"f2680ac0ae84f0980dbfb19842ec22005549507baa3436332fa16fa6776a06b2"} Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.002911 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2680ac0ae84f0980dbfb19842ec22005549507baa3436332fa16fa6776a06b2" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.003032 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-pq8r4" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.016381 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84b49964c9-d96qj" event={"ID":"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e","Type":"ContainerStarted","Data":"bb683c46344456c26442d9f75c79318f3cda06869c2bc3b68c7251729f08e3ae"} Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.578491 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8487496c64-cv9xr"] Feb 27 16:44:09 crc kubenswrapper[4814]: E0227 16:44:09.578866 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f5812c2-9952-4125-b055-dd2888fad3f5" containerName="placement-db-sync" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.578879 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f5812c2-9952-4125-b055-dd2888fad3f5" containerName="placement-db-sync" Feb 27 16:44:09 crc kubenswrapper[4814]: E0227 16:44:09.578905 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68c41339-6536-4998-901e-179aee23ac8c" containerName="oc" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.578912 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="68c41339-6536-4998-901e-179aee23ac8c" containerName="oc" Feb 27 16:44:09 crc kubenswrapper[4814]: E0227 16:44:09.578921 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3efc1a92-93e1-45a5-86c6-263ad2d4c36b" containerName="init" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.578929 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="3efc1a92-93e1-45a5-86c6-263ad2d4c36b" containerName="init" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.579096 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="3efc1a92-93e1-45a5-86c6-263ad2d4c36b" containerName="init" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.579113 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f5812c2-9952-4125-b055-dd2888fad3f5" containerName="placement-db-sync" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.579137 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="68c41339-6536-4998-901e-179aee23ac8c" containerName="oc" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.580074 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.592111 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.592562 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.592679 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-k25g2" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.592798 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.592899 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.604977 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8487496c64-cv9xr"] Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.708555 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-845d8f46cb-qhstj" podUID="6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.729928 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/13970ebf-9bd1-4292-a552-3f2d6bfa494c-internal-tls-certs\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.729970 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/13970ebf-9bd1-4292-a552-3f2d6bfa494c-public-tls-certs\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.730018 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13970ebf-9bd1-4292-a552-3f2d6bfa494c-logs\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.730104 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13970ebf-9bd1-4292-a552-3f2d6bfa494c-config-data\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.730152 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13970ebf-9bd1-4292-a552-3f2d6bfa494c-scripts\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.730181 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13970ebf-9bd1-4292-a552-3f2d6bfa494c-combined-ca-bundle\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.730363 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqg86\" (UniqueName: \"kubernetes.io/projected/13970ebf-9bd1-4292-a552-3f2d6bfa494c-kube-api-access-kqg86\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.771664 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-65784f76f6-f2pcp" podUID="8af960c4-8a04-42d8-83bf-9d03c23ad333" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.832015 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqg86\" (UniqueName: \"kubernetes.io/projected/13970ebf-9bd1-4292-a552-3f2d6bfa494c-kube-api-access-kqg86\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.832065 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/13970ebf-9bd1-4292-a552-3f2d6bfa494c-internal-tls-certs\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.832083 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/13970ebf-9bd1-4292-a552-3f2d6bfa494c-public-tls-certs\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.832124 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13970ebf-9bd1-4292-a552-3f2d6bfa494c-logs\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.832153 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13970ebf-9bd1-4292-a552-3f2d6bfa494c-config-data\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.832184 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13970ebf-9bd1-4292-a552-3f2d6bfa494c-scripts\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.832203 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13970ebf-9bd1-4292-a552-3f2d6bfa494c-combined-ca-bundle\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.833202 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13970ebf-9bd1-4292-a552-3f2d6bfa494c-logs\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.837991 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13970ebf-9bd1-4292-a552-3f2d6bfa494c-config-data\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.841886 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13970ebf-9bd1-4292-a552-3f2d6bfa494c-combined-ca-bundle\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.842240 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/13970ebf-9bd1-4292-a552-3f2d6bfa494c-public-tls-certs\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.850198 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/13970ebf-9bd1-4292-a552-3f2d6bfa494c-internal-tls-certs\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.854951 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqg86\" (UniqueName: \"kubernetes.io/projected/13970ebf-9bd1-4292-a552-3f2d6bfa494c-kube-api-access-kqg86\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.860758 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13970ebf-9bd1-4292-a552-3f2d6bfa494c-scripts\") pod \"placement-8487496c64-cv9xr\" (UID: \"13970ebf-9bd1-4292-a552-3f2d6bfa494c\") " pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:09 crc kubenswrapper[4814]: I0227 16:44:09.948534 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.065823 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84b49964c9-d96qj" event={"ID":"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e","Type":"ContainerStarted","Data":"331c051ddf16bda9435108936149a5fd5e9a8bf59c8ff3f0ea39c514bac30841"} Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.082651 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" event={"ID":"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b","Type":"ContainerStarted","Data":"68548c4ed8c1e2bfba57d66f743c2a956ecb0ae1075dfaaeaa83ce5e14ec6a5e"} Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.095888 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-84b49964c9-d96qj" podStartSLOduration=5.873437209 podStartE2EDuration="12.095869841s" podCreationTimestamp="2026-02-27 16:43:58 +0000 UTC" firstStartedPulling="2026-02-27 16:44:02.186964001 +0000 UTC m=+1254.639588831" lastFinishedPulling="2026-02-27 16:44:08.409396633 +0000 UTC m=+1260.862021463" observedRunningTime="2026-02-27 16:44:10.085751044 +0000 UTC m=+1262.538375874" watchObservedRunningTime="2026-02-27 16:44:10.095869841 +0000 UTC m=+1262.548494671" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.116904 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" podStartSLOduration=6.694416752 podStartE2EDuration="12.116888249s" podCreationTimestamp="2026-02-27 16:43:58 +0000 UTC" firstStartedPulling="2026-02-27 16:44:02.989123496 +0000 UTC m=+1255.441748326" lastFinishedPulling="2026-02-27 16:44:08.411594993 +0000 UTC m=+1260.864219823" observedRunningTime="2026-02-27 16:44:10.115684602 +0000 UTC m=+1262.568309432" watchObservedRunningTime="2026-02-27 16:44:10.116888249 +0000 UTC m=+1262.569513079" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.249825 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-786585dcfc-sm2x8"] Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.251532 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.281321 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6c5f495fc4-tkkrn"] Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.282770 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.297969 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-786585dcfc-sm2x8"] Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.319298 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6c5f495fc4-tkkrn"] Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.347905 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bbc2d20-3264-487d-8571-71bc22fba348-config-data-custom\") pod \"barbican-keystone-listener-6c5f495fc4-tkkrn\" (UID: \"2bbc2d20-3264-487d-8571-71bc22fba348\") " pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.347989 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b43b2496-ffec-4197-b0df-f03a3fc29ebf-config-data-custom\") pod \"barbican-worker-786585dcfc-sm2x8\" (UID: \"b43b2496-ffec-4197-b0df-f03a3fc29ebf\") " pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.348014 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b43b2496-ffec-4197-b0df-f03a3fc29ebf-config-data\") pod \"barbican-worker-786585dcfc-sm2x8\" (UID: \"b43b2496-ffec-4197-b0df-f03a3fc29ebf\") " pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.348036 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bbc2d20-3264-487d-8571-71bc22fba348-logs\") pod \"barbican-keystone-listener-6c5f495fc4-tkkrn\" (UID: \"2bbc2d20-3264-487d-8571-71bc22fba348\") " pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.348065 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbc2d20-3264-487d-8571-71bc22fba348-combined-ca-bundle\") pod \"barbican-keystone-listener-6c5f495fc4-tkkrn\" (UID: \"2bbc2d20-3264-487d-8571-71bc22fba348\") " pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.348114 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m7gn\" (UniqueName: \"kubernetes.io/projected/2bbc2d20-3264-487d-8571-71bc22fba348-kube-api-access-8m7gn\") pod \"barbican-keystone-listener-6c5f495fc4-tkkrn\" (UID: \"2bbc2d20-3264-487d-8571-71bc22fba348\") " pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.348129 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w84bc\" (UniqueName: \"kubernetes.io/projected/b43b2496-ffec-4197-b0df-f03a3fc29ebf-kube-api-access-w84bc\") pod \"barbican-worker-786585dcfc-sm2x8\" (UID: \"b43b2496-ffec-4197-b0df-f03a3fc29ebf\") " pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.348192 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bbc2d20-3264-487d-8571-71bc22fba348-config-data\") pod \"barbican-keystone-listener-6c5f495fc4-tkkrn\" (UID: \"2bbc2d20-3264-487d-8571-71bc22fba348\") " pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.348211 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b43b2496-ffec-4197-b0df-f03a3fc29ebf-combined-ca-bundle\") pod \"barbican-worker-786585dcfc-sm2x8\" (UID: \"b43b2496-ffec-4197-b0df-f03a3fc29ebf\") " pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.348225 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b43b2496-ffec-4197-b0df-f03a3fc29ebf-logs\") pod \"barbican-worker-786585dcfc-sm2x8\" (UID: \"b43b2496-ffec-4197-b0df-f03a3fc29ebf\") " pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.458635 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6f4999f464-wcs2l"] Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.458858 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6f4999f464-wcs2l" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api-log" containerID="cri-o://bf258e08576cf85838807e3706331bb269ce81b0024fdb537b74ec012819b4ba" gracePeriod=30 Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.459193 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6f4999f464-wcs2l" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api" containerID="cri-o://2421ada8cb323269d6db5f817fde7424f4db0ab7c760ab8bd82858da18be7108" gracePeriod=30 Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.463812 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b43b2496-ffec-4197-b0df-f03a3fc29ebf-config-data\") pod \"barbican-worker-786585dcfc-sm2x8\" (UID: \"b43b2496-ffec-4197-b0df-f03a3fc29ebf\") " pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.463856 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bbc2d20-3264-487d-8571-71bc22fba348-logs\") pod \"barbican-keystone-listener-6c5f495fc4-tkkrn\" (UID: \"2bbc2d20-3264-487d-8571-71bc22fba348\") " pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.463891 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbc2d20-3264-487d-8571-71bc22fba348-combined-ca-bundle\") pod \"barbican-keystone-listener-6c5f495fc4-tkkrn\" (UID: \"2bbc2d20-3264-487d-8571-71bc22fba348\") " pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.463939 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m7gn\" (UniqueName: \"kubernetes.io/projected/2bbc2d20-3264-487d-8571-71bc22fba348-kube-api-access-8m7gn\") pod \"barbican-keystone-listener-6c5f495fc4-tkkrn\" (UID: \"2bbc2d20-3264-487d-8571-71bc22fba348\") " pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.463953 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w84bc\" (UniqueName: \"kubernetes.io/projected/b43b2496-ffec-4197-b0df-f03a3fc29ebf-kube-api-access-w84bc\") pod \"barbican-worker-786585dcfc-sm2x8\" (UID: \"b43b2496-ffec-4197-b0df-f03a3fc29ebf\") " pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.464019 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bbc2d20-3264-487d-8571-71bc22fba348-config-data\") pod \"barbican-keystone-listener-6c5f495fc4-tkkrn\" (UID: \"2bbc2d20-3264-487d-8571-71bc22fba348\") " pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.464040 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b43b2496-ffec-4197-b0df-f03a3fc29ebf-combined-ca-bundle\") pod \"barbican-worker-786585dcfc-sm2x8\" (UID: \"b43b2496-ffec-4197-b0df-f03a3fc29ebf\") " pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.464055 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b43b2496-ffec-4197-b0df-f03a3fc29ebf-logs\") pod \"barbican-worker-786585dcfc-sm2x8\" (UID: \"b43b2496-ffec-4197-b0df-f03a3fc29ebf\") " pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.464082 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bbc2d20-3264-487d-8571-71bc22fba348-config-data-custom\") pod \"barbican-keystone-listener-6c5f495fc4-tkkrn\" (UID: \"2bbc2d20-3264-487d-8571-71bc22fba348\") " pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.464108 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b43b2496-ffec-4197-b0df-f03a3fc29ebf-config-data-custom\") pod \"barbican-worker-786585dcfc-sm2x8\" (UID: \"b43b2496-ffec-4197-b0df-f03a3fc29ebf\") " pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.465407 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bbc2d20-3264-487d-8571-71bc22fba348-logs\") pod \"barbican-keystone-listener-6c5f495fc4-tkkrn\" (UID: \"2bbc2d20-3264-487d-8571-71bc22fba348\") " pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.466904 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b43b2496-ffec-4197-b0df-f03a3fc29ebf-logs\") pod \"barbican-worker-786585dcfc-sm2x8\" (UID: \"b43b2496-ffec-4197-b0df-f03a3fc29ebf\") " pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.472776 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b43b2496-ffec-4197-b0df-f03a3fc29ebf-config-data\") pod \"barbican-worker-786585dcfc-sm2x8\" (UID: \"b43b2496-ffec-4197-b0df-f03a3fc29ebf\") " pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.476131 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbc2d20-3264-487d-8571-71bc22fba348-combined-ca-bundle\") pod \"barbican-keystone-listener-6c5f495fc4-tkkrn\" (UID: \"2bbc2d20-3264-487d-8571-71bc22fba348\") " pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.476627 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b43b2496-ffec-4197-b0df-f03a3fc29ebf-combined-ca-bundle\") pod \"barbican-worker-786585dcfc-sm2x8\" (UID: \"b43b2496-ffec-4197-b0df-f03a3fc29ebf\") " pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.477062 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bbc2d20-3264-487d-8571-71bc22fba348-config-data\") pod \"barbican-keystone-listener-6c5f495fc4-tkkrn\" (UID: \"2bbc2d20-3264-487d-8571-71bc22fba348\") " pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.479098 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bbc2d20-3264-487d-8571-71bc22fba348-config-data-custom\") pod \"barbican-keystone-listener-6c5f495fc4-tkkrn\" (UID: \"2bbc2d20-3264-487d-8571-71bc22fba348\") " pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.482483 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b43b2496-ffec-4197-b0df-f03a3fc29ebf-config-data-custom\") pod \"barbican-worker-786585dcfc-sm2x8\" (UID: \"b43b2496-ffec-4197-b0df-f03a3fc29ebf\") " pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.516872 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m7gn\" (UniqueName: \"kubernetes.io/projected/2bbc2d20-3264-487d-8571-71bc22fba348-kube-api-access-8m7gn\") pod \"barbican-keystone-listener-6c5f495fc4-tkkrn\" (UID: \"2bbc2d20-3264-487d-8571-71bc22fba348\") " pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.575791 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w84bc\" (UniqueName: \"kubernetes.io/projected/b43b2496-ffec-4197-b0df-f03a3fc29ebf-kube-api-access-w84bc\") pod \"barbican-worker-786585dcfc-sm2x8\" (UID: \"b43b2496-ffec-4197-b0df-f03a3fc29ebf\") " pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.581926 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6f4999f464-wcs2l" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.582245 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6f4999f464-wcs2l" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.619883 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f4999f464-wcs2l" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.620382 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f4999f464-wcs2l" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.643734 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-786585dcfc-sm2x8" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.658539 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa4535b0-2dac-45e8-9a9a-3ea58bb75658" path="/var/lib/kubelet/pods/aa4535b0-2dac-45e8-9a9a-3ea58bb75658/volumes" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.666467 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-f8687c5bd-zdnxs"] Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.668985 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-f8687c5bd-zdnxs"] Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.669156 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.671930 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.695334 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8487496c64-cv9xr"] Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.781323 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-internal-tls-certs\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.781747 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-config-data\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.781851 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk62l\" (UniqueName: \"kubernetes.io/projected/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-kube-api-access-wk62l\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.781879 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-config-data-custom\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.781905 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-public-tls-certs\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.781971 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-logs\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.782016 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-combined-ca-bundle\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.883394 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-internal-tls-certs\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.883453 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-config-data\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.883512 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk62l\" (UniqueName: \"kubernetes.io/projected/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-kube-api-access-wk62l\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.883537 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-config-data-custom\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.883555 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-public-tls-certs\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.883600 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-logs\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.883631 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-combined-ca-bundle\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.888950 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-logs\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.889130 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-combined-ca-bundle\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.891102 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-public-tls-certs\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.893792 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-internal-tls-certs\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.894465 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-config-data-custom\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.894733 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-config-data\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.919959 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk62l\" (UniqueName: \"kubernetes.io/projected/1121d7e8-a956-4a6b-929c-b1aecbaf16a4-kube-api-access-wk62l\") pod \"barbican-api-f8687c5bd-zdnxs\" (UID: \"1121d7e8-a956-4a6b-929c-b1aecbaf16a4\") " pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:10 crc kubenswrapper[4814]: I0227 16:44:10.999533 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:11 crc kubenswrapper[4814]: I0227 16:44:11.143825 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8487496c64-cv9xr" event={"ID":"13970ebf-9bd1-4292-a552-3f2d6bfa494c","Type":"ContainerStarted","Data":"b220553cbeda52305ff6a2dafb0d488843451fa47f8ff3283874cd9f35226f12"} Feb 27 16:44:11 crc kubenswrapper[4814]: I0227 16:44:11.143875 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8487496c64-cv9xr" event={"ID":"13970ebf-9bd1-4292-a552-3f2d6bfa494c","Type":"ContainerStarted","Data":"97d8cd5b6c51e8d9f8e1866ac75ed120494045db327b931776c519c585985fa6"} Feb 27 16:44:11 crc kubenswrapper[4814]: I0227 16:44:11.145506 4814 generic.go:334] "Generic (PLEG): container finished" podID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerID="bf258e08576cf85838807e3706331bb269ce81b0024fdb537b74ec012819b4ba" exitCode=143 Feb 27 16:44:11 crc kubenswrapper[4814]: I0227 16:44:11.145564 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f4999f464-wcs2l" event={"ID":"c0b8557d-2397-4c7d-b6ab-6044db5f0129","Type":"ContainerDied","Data":"bf258e08576cf85838807e3706331bb269ce81b0024fdb537b74ec012819b4ba"} Feb 27 16:44:11 crc kubenswrapper[4814]: I0227 16:44:11.174966 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-786585dcfc-sm2x8"] Feb 27 16:44:11 crc kubenswrapper[4814]: I0227 16:44:11.216912 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6c5f495fc4-tkkrn"] Feb 27 16:44:11 crc kubenswrapper[4814]: I0227 16:44:11.339875 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-f8687c5bd-zdnxs"] Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.163217 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f8687c5bd-zdnxs" event={"ID":"1121d7e8-a956-4a6b-929c-b1aecbaf16a4","Type":"ContainerStarted","Data":"83de672a9f6e70c33bdcf10942d40ebb826b9a3847a21cd021b01f37ca8a8081"} Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.163505 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f8687c5bd-zdnxs" event={"ID":"1121d7e8-a956-4a6b-929c-b1aecbaf16a4","Type":"ContainerStarted","Data":"19d3caf2b2a3114978b1dc9c4f15e8516a465afad64a8e0bf5e4cadb41207965"} Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.170189 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-786585dcfc-sm2x8" event={"ID":"b43b2496-ffec-4197-b0df-f03a3fc29ebf","Type":"ContainerStarted","Data":"2b6fe8f4211c42264b0c678c3215e8b76f0286685064f8132a18dbaf1a4bf789"} Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.170229 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-786585dcfc-sm2x8" event={"ID":"b43b2496-ffec-4197-b0df-f03a3fc29ebf","Type":"ContainerStarted","Data":"fbb0bd1e5cae6ced2be7c609c6ffad5ade464686dc0076a0536df4298ddc2b30"} Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.170243 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-786585dcfc-sm2x8" event={"ID":"b43b2496-ffec-4197-b0df-f03a3fc29ebf","Type":"ContainerStarted","Data":"d261a6424bab8a4699a54afcb315eb9737ccbb0e8a0524d2898bc2807332c247"} Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.174009 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8487496c64-cv9xr" event={"ID":"13970ebf-9bd1-4292-a552-3f2d6bfa494c","Type":"ContainerStarted","Data":"53e95a10d36ca8f972e13fc2e671a9e95d0867388810a05978dafb3fe518d915"} Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.174205 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.174244 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.175778 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" event={"ID":"2bbc2d20-3264-487d-8571-71bc22fba348","Type":"ContainerStarted","Data":"1398a407f31f8daec851e8e690e3d652e6e7a27b5cdb0775faf05788067cc5de"} Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.175820 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" event={"ID":"2bbc2d20-3264-487d-8571-71bc22fba348","Type":"ContainerStarted","Data":"fe488b47b9f3fde5260a2409b392fe751ddd27725201d268182f60e46ec21d6a"} Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.175832 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" event={"ID":"2bbc2d20-3264-487d-8571-71bc22fba348","Type":"ContainerStarted","Data":"e13630f8412e21077df89deccfa4b716cbf9cdaba05699c0ac69969441ce207b"} Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.189652 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-786585dcfc-sm2x8" podStartSLOduration=2.18963508 podStartE2EDuration="2.18963508s" podCreationTimestamp="2026-02-27 16:44:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:12.184280862 +0000 UTC m=+1264.636905692" watchObservedRunningTime="2026-02-27 16:44:12.18963508 +0000 UTC m=+1264.642259910" Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.228135 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-8487496c64-cv9xr" podStartSLOduration=3.228117676 podStartE2EDuration="3.228117676s" podCreationTimestamp="2026-02-27 16:44:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:12.220079835 +0000 UTC m=+1264.672704665" watchObservedRunningTime="2026-02-27 16:44:12.228117676 +0000 UTC m=+1264.680742506" Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.244274 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-84b49964c9-d96qj"] Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.244472 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-84b49964c9-d96qj" podUID="d34c8c2a-0bf5-43a8-9ad9-110875e49b3e" containerName="barbican-worker-log" containerID="cri-o://bb683c46344456c26442d9f75c79318f3cda06869c2bc3b68c7251729f08e3ae" gracePeriod=30 Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.244843 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-84b49964c9-d96qj" podUID="d34c8c2a-0bf5-43a8-9ad9-110875e49b3e" containerName="barbican-worker" containerID="cri-o://331c051ddf16bda9435108936149a5fd5e9a8bf59c8ff3f0ea39c514bac30841" gracePeriod=30 Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.252333 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6c5f495fc4-tkkrn" podStartSLOduration=2.252317186 podStartE2EDuration="2.252317186s" podCreationTimestamp="2026-02-27 16:44:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:12.242910561 +0000 UTC m=+1264.695535391" watchObservedRunningTime="2026-02-27 16:44:12.252317186 +0000 UTC m=+1264.704942016" Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.303910 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-66c4c86998-k7mxw"] Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.304202 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" podUID="6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b" containerName="barbican-keystone-listener-log" containerID="cri-o://6a28d6cd48600190d077b9746978ba0bd442982fe85401efd2b8470fe849af70" gracePeriod=30 Feb 27 16:44:12 crc kubenswrapper[4814]: I0227 16:44:12.304650 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" podUID="6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b" containerName="barbican-keystone-listener" containerID="cri-o://68548c4ed8c1e2bfba57d66f743c2a956ecb0ae1075dfaaeaa83ce5e14ec6a5e" gracePeriod=30 Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.191346 4814 generic.go:334] "Generic (PLEG): container finished" podID="6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b" containerID="6a28d6cd48600190d077b9746978ba0bd442982fe85401efd2b8470fe849af70" exitCode=143 Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.191440 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" event={"ID":"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b","Type":"ContainerDied","Data":"6a28d6cd48600190d077b9746978ba0bd442982fe85401efd2b8470fe849af70"} Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.195396 4814 generic.go:334] "Generic (PLEG): container finished" podID="d34c8c2a-0bf5-43a8-9ad9-110875e49b3e" containerID="331c051ddf16bda9435108936149a5fd5e9a8bf59c8ff3f0ea39c514bac30841" exitCode=0 Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.195426 4814 generic.go:334] "Generic (PLEG): container finished" podID="d34c8c2a-0bf5-43a8-9ad9-110875e49b3e" containerID="bb683c46344456c26442d9f75c79318f3cda06869c2bc3b68c7251729f08e3ae" exitCode=143 Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.195452 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84b49964c9-d96qj" event={"ID":"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e","Type":"ContainerDied","Data":"331c051ddf16bda9435108936149a5fd5e9a8bf59c8ff3f0ea39c514bac30841"} Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.195488 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84b49964c9-d96qj" event={"ID":"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e","Type":"ContainerDied","Data":"bb683c46344456c26442d9f75c79318f3cda06869c2bc3b68c7251729f08e3ae"} Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.198124 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f8687c5bd-zdnxs" event={"ID":"1121d7e8-a956-4a6b-929c-b1aecbaf16a4","Type":"ContainerStarted","Data":"dbd8e9140258f9ce6df10845f7bb8c4cffb6cf7a2369a970b0592bc5240518f3"} Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.229324 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-f8687c5bd-zdnxs" podStartSLOduration=3.229306284 podStartE2EDuration="3.229306284s" podCreationTimestamp="2026-02-27 16:44:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:13.225423342 +0000 UTC m=+1265.678048172" watchObservedRunningTime="2026-02-27 16:44:13.229306284 +0000 UTC m=+1265.681931124" Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.850467 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.856654 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-combined-ca-bundle\") pod \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.856767 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-logs\") pod \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.856860 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2f8l\" (UniqueName: \"kubernetes.io/projected/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-kube-api-access-q2f8l\") pod \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.856951 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-config-data\") pod \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.856999 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-config-data-custom\") pod \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\" (UID: \"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e\") " Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.857163 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-logs" (OuterVolumeSpecName: "logs") pod "d34c8c2a-0bf5-43a8-9ad9-110875e49b3e" (UID: "d34c8c2a-0bf5-43a8-9ad9-110875e49b3e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.857397 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.872965 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d34c8c2a-0bf5-43a8-9ad9-110875e49b3e" (UID: "d34c8c2a-0bf5-43a8-9ad9-110875e49b3e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.877552 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-kube-api-access-q2f8l" (OuterVolumeSpecName: "kube-api-access-q2f8l") pod "d34c8c2a-0bf5-43a8-9ad9-110875e49b3e" (UID: "d34c8c2a-0bf5-43a8-9ad9-110875e49b3e"). InnerVolumeSpecName "kube-api-access-q2f8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.947514 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d34c8c2a-0bf5-43a8-9ad9-110875e49b3e" (UID: "d34c8c2a-0bf5-43a8-9ad9-110875e49b3e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.959111 4814 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.959144 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.959153 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2f8l\" (UniqueName: \"kubernetes.io/projected/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-kube-api-access-q2f8l\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:13 crc kubenswrapper[4814]: I0227 16:44:13.972390 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-config-data" (OuterVolumeSpecName: "config-data") pod "d34c8c2a-0bf5-43a8-9ad9-110875e49b3e" (UID: "d34c8c2a-0bf5-43a8-9ad9-110875e49b3e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.061131 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.136332 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.163803 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-logs\") pod \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.163897 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmsrk\" (UniqueName: \"kubernetes.io/projected/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-kube-api-access-mmsrk\") pod \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.166519 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-config-data\") pod \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.166629 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-config-data-custom\") pod \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.166759 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-combined-ca-bundle\") pod \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\" (UID: \"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b\") " Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.167441 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-logs" (OuterVolumeSpecName: "logs") pod "6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b" (UID: "6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.170520 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b" (UID: "6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.174403 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-kube-api-access-mmsrk" (OuterVolumeSpecName: "kube-api-access-mmsrk") pod "6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b" (UID: "6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b"). InnerVolumeSpecName "kube-api-access-mmsrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.177591 4814 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.177622 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.177634 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmsrk\" (UniqueName: \"kubernetes.io/projected/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-kube-api-access-mmsrk\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.196364 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b" (UID: "6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.209432 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84b49964c9-d96qj" event={"ID":"d34c8c2a-0bf5-43a8-9ad9-110875e49b3e","Type":"ContainerDied","Data":"459674b1b87d9fd9bbd1c8e019783ce1064ec0a6b2352851c8b1e64c46e23614"} Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.209501 4814 scope.go:117] "RemoveContainer" containerID="331c051ddf16bda9435108936149a5fd5e9a8bf59c8ff3f0ea39c514bac30841" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.209636 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-84b49964c9-d96qj" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.214680 4814 generic.go:334] "Generic (PLEG): container finished" podID="6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b" containerID="68548c4ed8c1e2bfba57d66f743c2a956ecb0ae1075dfaaeaa83ce5e14ec6a5e" exitCode=0 Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.215488 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.216134 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" event={"ID":"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b","Type":"ContainerDied","Data":"68548c4ed8c1e2bfba57d66f743c2a956ecb0ae1075dfaaeaa83ce5e14ec6a5e"} Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.216167 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.216178 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-66c4c86998-k7mxw" event={"ID":"6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b","Type":"ContainerDied","Data":"0dbb3beb7b5d2c86451925d2da96ad6882e8810a5a85fcd89a359e636134459e"} Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.216215 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.228794 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.275175 4814 scope.go:117] "RemoveContainer" containerID="bb683c46344456c26442d9f75c79318f3cda06869c2bc3b68c7251729f08e3ae" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.280574 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.286554 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-config-data" (OuterVolumeSpecName: "config-data") pod "6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b" (UID: "6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.309399 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-84b49964c9-d96qj"] Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.319166 4814 scope.go:117] "RemoveContainer" containerID="68548c4ed8c1e2bfba57d66f743c2a956ecb0ae1075dfaaeaa83ce5e14ec6a5e" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.322432 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-84b49964c9-d96qj"] Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.336904 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-rv8p6"] Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.337158 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" podUID="93476b12-7c88-4276-b58f-2a6173a7eca1" containerName="dnsmasq-dns" containerID="cri-o://3ca9f51e66aecebea6995d1f41a8b77d3b1666e0d062853132be7776f7561b7d" gracePeriod=10 Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.383146 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.407125 4814 scope.go:117] "RemoveContainer" containerID="6a28d6cd48600190d077b9746978ba0bd442982fe85401efd2b8470fe849af70" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.469021 4814 scope.go:117] "RemoveContainer" containerID="68548c4ed8c1e2bfba57d66f743c2a956ecb0ae1075dfaaeaa83ce5e14ec6a5e" Feb 27 16:44:14 crc kubenswrapper[4814]: E0227 16:44:14.475659 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68548c4ed8c1e2bfba57d66f743c2a956ecb0ae1075dfaaeaa83ce5e14ec6a5e\": container with ID starting with 68548c4ed8c1e2bfba57d66f743c2a956ecb0ae1075dfaaeaa83ce5e14ec6a5e not found: ID does not exist" containerID="68548c4ed8c1e2bfba57d66f743c2a956ecb0ae1075dfaaeaa83ce5e14ec6a5e" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.475704 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68548c4ed8c1e2bfba57d66f743c2a956ecb0ae1075dfaaeaa83ce5e14ec6a5e"} err="failed to get container status \"68548c4ed8c1e2bfba57d66f743c2a956ecb0ae1075dfaaeaa83ce5e14ec6a5e\": rpc error: code = NotFound desc = could not find container \"68548c4ed8c1e2bfba57d66f743c2a956ecb0ae1075dfaaeaa83ce5e14ec6a5e\": container with ID starting with 68548c4ed8c1e2bfba57d66f743c2a956ecb0ae1075dfaaeaa83ce5e14ec6a5e not found: ID does not exist" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.475728 4814 scope.go:117] "RemoveContainer" containerID="6a28d6cd48600190d077b9746978ba0bd442982fe85401efd2b8470fe849af70" Feb 27 16:44:14 crc kubenswrapper[4814]: E0227 16:44:14.479749 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a28d6cd48600190d077b9746978ba0bd442982fe85401efd2b8470fe849af70\": container with ID starting with 6a28d6cd48600190d077b9746978ba0bd442982fe85401efd2b8470fe849af70 not found: ID does not exist" containerID="6a28d6cd48600190d077b9746978ba0bd442982fe85401efd2b8470fe849af70" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.479791 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a28d6cd48600190d077b9746978ba0bd442982fe85401efd2b8470fe849af70"} err="failed to get container status \"6a28d6cd48600190d077b9746978ba0bd442982fe85401efd2b8470fe849af70\": rpc error: code = NotFound desc = could not find container \"6a28d6cd48600190d077b9746978ba0bd442982fe85401efd2b8470fe849af70\": container with ID starting with 6a28d6cd48600190d077b9746978ba0bd442982fe85401efd2b8470fe849af70 not found: ID does not exist" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.509549 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d34c8c2a-0bf5-43a8-9ad9-110875e49b3e" path="/var/lib/kubelet/pods/d34c8c2a-0bf5-43a8-9ad9-110875e49b3e/volumes" Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.592062 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-66c4c86998-k7mxw"] Feb 27 16:44:14 crc kubenswrapper[4814]: I0227 16:44:14.603300 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-66c4c86998-k7mxw"] Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.239749 4814 generic.go:334] "Generic (PLEG): container finished" podID="7bcf9bc6-1fb4-47b4-b12d-017902e4992e" containerID="5c90611e97642d06afcbae7dc9e52e2df50a2e1faaa225ef508016d671dd756f" exitCode=0 Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.240002 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vqkl5" event={"ID":"7bcf9bc6-1fb4-47b4-b12d-017902e4992e","Type":"ContainerDied","Data":"5c90611e97642d06afcbae7dc9e52e2df50a2e1faaa225ef508016d671dd756f"} Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.245688 4814 generic.go:334] "Generic (PLEG): container finished" podID="93476b12-7c88-4276-b58f-2a6173a7eca1" containerID="3ca9f51e66aecebea6995d1f41a8b77d3b1666e0d062853132be7776f7561b7d" exitCode=0 Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.246827 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" event={"ID":"93476b12-7c88-4276-b58f-2a6173a7eca1","Type":"ContainerDied","Data":"3ca9f51e66aecebea6995d1f41a8b77d3b1666e0d062853132be7776f7561b7d"} Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.429580 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.503665 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-dns-swift-storage-0\") pod \"93476b12-7c88-4276-b58f-2a6173a7eca1\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.503736 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-ovsdbserver-nb\") pod \"93476b12-7c88-4276-b58f-2a6173a7eca1\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.503791 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-config\") pod \"93476b12-7c88-4276-b58f-2a6173a7eca1\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.503838 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-ovsdbserver-sb\") pod \"93476b12-7c88-4276-b58f-2a6173a7eca1\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.503974 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-dns-svc\") pod \"93476b12-7c88-4276-b58f-2a6173a7eca1\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.504009 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd7n5\" (UniqueName: \"kubernetes.io/projected/93476b12-7c88-4276-b58f-2a6173a7eca1-kube-api-access-vd7n5\") pod \"93476b12-7c88-4276-b58f-2a6173a7eca1\" (UID: \"93476b12-7c88-4276-b58f-2a6173a7eca1\") " Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.520365 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93476b12-7c88-4276-b58f-2a6173a7eca1-kube-api-access-vd7n5" (OuterVolumeSpecName: "kube-api-access-vd7n5") pod "93476b12-7c88-4276-b58f-2a6173a7eca1" (UID: "93476b12-7c88-4276-b58f-2a6173a7eca1"). InnerVolumeSpecName "kube-api-access-vd7n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.594588 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "93476b12-7c88-4276-b58f-2a6173a7eca1" (UID: "93476b12-7c88-4276-b58f-2a6173a7eca1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.609520 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd7n5\" (UniqueName: \"kubernetes.io/projected/93476b12-7c88-4276-b58f-2a6173a7eca1-kube-api-access-vd7n5\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.609555 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.620623 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "93476b12-7c88-4276-b58f-2a6173a7eca1" (UID: "93476b12-7c88-4276-b58f-2a6173a7eca1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.624636 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-config" (OuterVolumeSpecName: "config") pod "93476b12-7c88-4276-b58f-2a6173a7eca1" (UID: "93476b12-7c88-4276-b58f-2a6173a7eca1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.643092 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "93476b12-7c88-4276-b58f-2a6173a7eca1" (UID: "93476b12-7c88-4276-b58f-2a6173a7eca1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.676243 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "93476b12-7c88-4276-b58f-2a6173a7eca1" (UID: "93476b12-7c88-4276-b58f-2a6173a7eca1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.702414 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f4999f464-wcs2l" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.702719 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f4999f464-wcs2l" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.710984 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.711011 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.711022 4814 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.711030 4814 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93476b12-7c88-4276-b58f-2a6173a7eca1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:15 crc kubenswrapper[4814]: I0227 16:44:15.739206 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:16 crc kubenswrapper[4814]: I0227 16:44:16.005075 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:16 crc kubenswrapper[4814]: I0227 16:44:16.149659 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f4999f464-wcs2l" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:50974->10.217.0.160:9311: read: connection reset by peer" Feb 27 16:44:16 crc kubenswrapper[4814]: I0227 16:44:16.149721 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f4999f464-wcs2l" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:50980->10.217.0.160:9311: read: connection reset by peer" Feb 27 16:44:16 crc kubenswrapper[4814]: I0227 16:44:16.314567 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" event={"ID":"93476b12-7c88-4276-b58f-2a6173a7eca1","Type":"ContainerDied","Data":"e130fb1fd885ffe92672fe9e9b0ee79f8573349034d6bb3a9ec9216a83215fb0"} Feb 27 16:44:16 crc kubenswrapper[4814]: I0227 16:44:16.314635 4814 scope.go:117] "RemoveContainer" containerID="3ca9f51e66aecebea6995d1f41a8b77d3b1666e0d062853132be7776f7561b7d" Feb 27 16:44:16 crc kubenswrapper[4814]: I0227 16:44:16.314798 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-rv8p6" Feb 27 16:44:16 crc kubenswrapper[4814]: I0227 16:44:16.343595 4814 generic.go:334] "Generic (PLEG): container finished" podID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerID="2421ada8cb323269d6db5f817fde7424f4db0ab7c760ab8bd82858da18be7108" exitCode=0 Feb 27 16:44:16 crc kubenswrapper[4814]: I0227 16:44:16.345109 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f4999f464-wcs2l" event={"ID":"c0b8557d-2397-4c7d-b6ab-6044db5f0129","Type":"ContainerDied","Data":"2421ada8cb323269d6db5f817fde7424f4db0ab7c760ab8bd82858da18be7108"} Feb 27 16:44:16 crc kubenswrapper[4814]: I0227 16:44:16.369326 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-rv8p6"] Feb 27 16:44:16 crc kubenswrapper[4814]: I0227 16:44:16.380695 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-rv8p6"] Feb 27 16:44:16 crc kubenswrapper[4814]: I0227 16:44:16.504347 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b" path="/var/lib/kubelet/pods/6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b/volumes" Feb 27 16:44:16 crc kubenswrapper[4814]: I0227 16:44:16.504954 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93476b12-7c88-4276-b58f-2a6173a7eca1" path="/var/lib/kubelet/pods/93476b12-7c88-4276-b58f-2a6173a7eca1/volumes" Feb 27 16:44:17 crc kubenswrapper[4814]: I0227 16:44:17.703727 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:19 crc kubenswrapper[4814]: I0227 16:44:19.058212 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-f8687c5bd-zdnxs" Feb 27 16:44:19 crc kubenswrapper[4814]: I0227 16:44:19.159442 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-855bdc9fdb-m79h8"] Feb 27 16:44:19 crc kubenswrapper[4814]: I0227 16:44:19.159662 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-855bdc9fdb-m79h8" podUID="3584eaa9-9218-4dca-848c-1781f316e4a2" containerName="barbican-api-log" containerID="cri-o://d2f2e337a2622ede196d96b51299a1cb5497818091db374683707fd677c69485" gracePeriod=30 Feb 27 16:44:19 crc kubenswrapper[4814]: I0227 16:44:19.159824 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-855bdc9fdb-m79h8" podUID="3584eaa9-9218-4dca-848c-1781f316e4a2" containerName="barbican-api" containerID="cri-o://f36a811f0c7b74a3b41ec489863d52d12d465987224d9968a5e7b15e60acf8ec" gracePeriod=30 Feb 27 16:44:19 crc kubenswrapper[4814]: I0227 16:44:19.203828 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f4999f464-wcs2l" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": dial tcp 10.217.0.160:9311: connect: connection refused" Feb 27 16:44:19 crc kubenswrapper[4814]: I0227 16:44:19.204075 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f4999f464-wcs2l" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": dial tcp 10.217.0.160:9311: connect: connection refused" Feb 27 16:44:19 crc kubenswrapper[4814]: I0227 16:44:19.372901 4814 generic.go:334] "Generic (PLEG): container finished" podID="3584eaa9-9218-4dca-848c-1781f316e4a2" containerID="d2f2e337a2622ede196d96b51299a1cb5497818091db374683707fd677c69485" exitCode=143 Feb 27 16:44:19 crc kubenswrapper[4814]: I0227 16:44:19.372936 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-855bdc9fdb-m79h8" event={"ID":"3584eaa9-9218-4dca-848c-1781f316e4a2","Type":"ContainerDied","Data":"d2f2e337a2622ede196d96b51299a1cb5497818091db374683707fd677c69485"} Feb 27 16:44:19 crc kubenswrapper[4814]: I0227 16:44:19.671783 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-845d8f46cb-qhstj" podUID="6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 27 16:44:19 crc kubenswrapper[4814]: I0227 16:44:19.764133 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-65784f76f6-f2pcp" podUID="8af960c4-8a04-42d8-83bf-9d03c23ad333" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.069503 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.174957 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-config-data\") pod \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.175273 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-etc-machine-id\") pod \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.175297 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-scripts\") pod \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.175318 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-db-sync-config-data\") pod \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.175371 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-combined-ca-bundle\") pod \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.175389 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpmlb\" (UniqueName: \"kubernetes.io/projected/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-kube-api-access-rpmlb\") pod \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\" (UID: \"7bcf9bc6-1fb4-47b4-b12d-017902e4992e\") " Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.181757 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-scripts" (OuterVolumeSpecName: "scripts") pod "7bcf9bc6-1fb4-47b4-b12d-017902e4992e" (UID: "7bcf9bc6-1fb4-47b4-b12d-017902e4992e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.182292 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7bcf9bc6-1fb4-47b4-b12d-017902e4992e" (UID: "7bcf9bc6-1fb4-47b4-b12d-017902e4992e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.187753 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-kube-api-access-rpmlb" (OuterVolumeSpecName: "kube-api-access-rpmlb") pod "7bcf9bc6-1fb4-47b4-b12d-017902e4992e" (UID: "7bcf9bc6-1fb4-47b4-b12d-017902e4992e"). InnerVolumeSpecName "kube-api-access-rpmlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.190404 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7bcf9bc6-1fb4-47b4-b12d-017902e4992e" (UID: "7bcf9bc6-1fb4-47b4-b12d-017902e4992e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.244430 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7bcf9bc6-1fb4-47b4-b12d-017902e4992e" (UID: "7bcf9bc6-1fb4-47b4-b12d-017902e4992e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.263745 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-config-data" (OuterVolumeSpecName: "config-data") pod "7bcf9bc6-1fb4-47b4-b12d-017902e4992e" (UID: "7bcf9bc6-1fb4-47b4-b12d-017902e4992e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.277202 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.277232 4814 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.277243 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.277269 4814 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.277281 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.277289 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpmlb\" (UniqueName: \"kubernetes.io/projected/7bcf9bc6-1fb4-47b4-b12d-017902e4992e-kube-api-access-rpmlb\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.404927 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vqkl5" event={"ID":"7bcf9bc6-1fb4-47b4-b12d-017902e4992e","Type":"ContainerDied","Data":"6ae9247a5eb96b5778623b9ad7c9cdb98e1e976b78d4caa4487eba75ce461606"} Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.404959 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ae9247a5eb96b5778623b9ad7c9cdb98e1e976b78d4caa4487eba75ce461606" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.405016 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vqkl5" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.595230 4814 scope.go:117] "RemoveContainer" containerID="ade4976a96e606bf2f3a5d54c59a659087beb45bf3c28e6f44af862ce1245604" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.613565 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.682678 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-config-data\") pod \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.682793 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-config-data-custom\") pod \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.682950 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jv95z\" (UniqueName: \"kubernetes.io/projected/c0b8557d-2397-4c7d-b6ab-6044db5f0129-kube-api-access-jv95z\") pod \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.683083 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-combined-ca-bundle\") pod \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.683199 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0b8557d-2397-4c7d-b6ab-6044db5f0129-logs\") pod \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\" (UID: \"c0b8557d-2397-4c7d-b6ab-6044db5f0129\") " Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.683697 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0b8557d-2397-4c7d-b6ab-6044db5f0129-logs" (OuterVolumeSpecName: "logs") pod "c0b8557d-2397-4c7d-b6ab-6044db5f0129" (UID: "c0b8557d-2397-4c7d-b6ab-6044db5f0129"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.687544 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c0b8557d-2397-4c7d-b6ab-6044db5f0129" (UID: "c0b8557d-2397-4c7d-b6ab-6044db5f0129"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.687878 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0b8557d-2397-4c7d-b6ab-6044db5f0129-kube-api-access-jv95z" (OuterVolumeSpecName: "kube-api-access-jv95z") pod "c0b8557d-2397-4c7d-b6ab-6044db5f0129" (UID: "c0b8557d-2397-4c7d-b6ab-6044db5f0129"). InnerVolumeSpecName "kube-api-access-jv95z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.725129 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0b8557d-2397-4c7d-b6ab-6044db5f0129" (UID: "c0b8557d-2397-4c7d-b6ab-6044db5f0129"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.755741 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-config-data" (OuterVolumeSpecName: "config-data") pod "c0b8557d-2397-4c7d-b6ab-6044db5f0129" (UID: "c0b8557d-2397-4c7d-b6ab-6044db5f0129"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.787654 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.787693 4814 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.787711 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jv95z\" (UniqueName: \"kubernetes.io/projected/c0b8557d-2397-4c7d-b6ab-6044db5f0129-kube-api-access-jv95z\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.787723 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0b8557d-2397-4c7d-b6ab-6044db5f0129-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:21 crc kubenswrapper[4814]: I0227 16:44:21.787734 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0b8557d-2397-4c7d-b6ab-6044db5f0129-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:21 crc kubenswrapper[4814]: E0227 16:44:21.871746 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.388955 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 16:44:22 crc kubenswrapper[4814]: E0227 16:44:22.389757 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bcf9bc6-1fb4-47b4-b12d-017902e4992e" containerName="cinder-db-sync" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.389774 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bcf9bc6-1fb4-47b4-b12d-017902e4992e" containerName="cinder-db-sync" Feb 27 16:44:22 crc kubenswrapper[4814]: E0227 16:44:22.389794 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api-log" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.389802 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api-log" Feb 27 16:44:22 crc kubenswrapper[4814]: E0227 16:44:22.389813 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93476b12-7c88-4276-b58f-2a6173a7eca1" containerName="dnsmasq-dns" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.389821 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="93476b12-7c88-4276-b58f-2a6173a7eca1" containerName="dnsmasq-dns" Feb 27 16:44:22 crc kubenswrapper[4814]: E0227 16:44:22.389833 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d34c8c2a-0bf5-43a8-9ad9-110875e49b3e" containerName="barbican-worker" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.389839 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="d34c8c2a-0bf5-43a8-9ad9-110875e49b3e" containerName="barbican-worker" Feb 27 16:44:22 crc kubenswrapper[4814]: E0227 16:44:22.389848 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d34c8c2a-0bf5-43a8-9ad9-110875e49b3e" containerName="barbican-worker-log" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.389855 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="d34c8c2a-0bf5-43a8-9ad9-110875e49b3e" containerName="barbican-worker-log" Feb 27 16:44:22 crc kubenswrapper[4814]: E0227 16:44:22.389868 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.389874 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api" Feb 27 16:44:22 crc kubenswrapper[4814]: E0227 16:44:22.389896 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93476b12-7c88-4276-b58f-2a6173a7eca1" containerName="init" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.389903 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="93476b12-7c88-4276-b58f-2a6173a7eca1" containerName="init" Feb 27 16:44:22 crc kubenswrapper[4814]: E0227 16:44:22.389912 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b" containerName="barbican-keystone-listener" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.389919 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b" containerName="barbican-keystone-listener" Feb 27 16:44:22 crc kubenswrapper[4814]: E0227 16:44:22.389938 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b" containerName="barbican-keystone-listener-log" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.390023 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b" containerName="barbican-keystone-listener-log" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.397928 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b" containerName="barbican-keystone-listener-log" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.397983 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="d34c8c2a-0bf5-43a8-9ad9-110875e49b3e" containerName="barbican-worker" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.397997 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ed6c6b6-db09-4f3a-ba22-c2b4a396e66b" containerName="barbican-keystone-listener" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.398015 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="93476b12-7c88-4276-b58f-2a6173a7eca1" containerName="dnsmasq-dns" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.398024 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.398039 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="d34c8c2a-0bf5-43a8-9ad9-110875e49b3e" containerName="barbican-worker-log" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.398051 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bcf9bc6-1fb4-47b4-b12d-017902e4992e" containerName="cinder-db-sync" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.398066 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" containerName="barbican-api-log" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.400137 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.406361 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.410058 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.410125 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.410168 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-89btm" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.412181 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.432171 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f4999f464-wcs2l" event={"ID":"c0b8557d-2397-4c7d-b6ab-6044db5f0129","Type":"ContainerDied","Data":"8a61d2ec28d34e888e8723543aaf09c8b1750abf0dcc6b37ec5dd3c7870e764c"} Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.432231 4814 scope.go:117] "RemoveContainer" containerID="2421ada8cb323269d6db5f817fde7424f4db0ab7c760ab8bd82858da18be7108" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.432401 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f4999f464-wcs2l" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.443688 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01245f0c-38fe-4f9a-93e5-c5c5d5f32259","Type":"ContainerStarted","Data":"146c26d9df94e91bd53b52abca4f6443179c2c8e2365581e20b12612c0129aca"} Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.443927 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" containerName="ceilometer-notification-agent" containerID="cri-o://c52d992512b39f9030e3eb2b20244de9ced5f2493ef47f17365e81f5d87b6cf2" gracePeriod=30 Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.444017 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.444037 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" containerName="proxy-httpd" containerID="cri-o://146c26d9df94e91bd53b52abca4f6443179c2c8e2365581e20b12612c0129aca" gracePeriod=30 Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.444099 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" containerName="sg-core" containerID="cri-o://fe992a34a5bd33198fa76c1d351deb8c6a8cb1d10babbc60cf65c741734db762" gracePeriod=30 Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.503269 4814 scope.go:117] "RemoveContainer" containerID="bf258e08576cf85838807e3706331bb269ce81b0024fdb537b74ec012819b4ba" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.521925 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-config-data\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.522125 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.522263 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-scripts\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.522381 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.522513 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlfrc\" (UniqueName: \"kubernetes.io/projected/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-kube-api-access-qlfrc\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.522597 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.574992 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-855bdc9fdb-m79h8" podUID="3584eaa9-9218-4dca-848c-1781f316e4a2" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.163:9311/healthcheck\": dial tcp 10.217.0.163:9311: connect: connection refused" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.575335 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-855bdc9fdb-m79h8" podUID="3584eaa9-9218-4dca-848c-1781f316e4a2" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.163:9311/healthcheck\": dial tcp 10.217.0.163:9311: connect: connection refused" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.584592 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-frg9p"] Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.598534 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.599369 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-frg9p"] Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.624599 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlfrc\" (UniqueName: \"kubernetes.io/projected/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-kube-api-access-qlfrc\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.624666 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.624737 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-config-data\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.624776 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.624838 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-scripts\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.624901 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.625040 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.639524 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6f4999f464-wcs2l"] Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.641329 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.645201 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlfrc\" (UniqueName: \"kubernetes.io/projected/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-kube-api-access-qlfrc\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.645585 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.653243 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-config-data\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.657482 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-scripts\") pod \"cinder-scheduler-0\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.663083 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6f4999f464-wcs2l"] Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.703183 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.704860 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.713881 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.721687 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.729928 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.729965 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.729985 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj995\" (UniqueName: \"kubernetes.io/projected/0f95f00d-d69f-4734-ba33-436879200784-kube-api-access-nj995\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.730094 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-config\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.730117 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-dns-svc\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.730141 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.759167 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.831847 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/448fcb03-9b59-4e38-8818-411c7c2dfcf0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.832250 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.832342 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.832375 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.832396 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-config-data\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.832424 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.832455 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj995\" (UniqueName: \"kubernetes.io/projected/0f95f00d-d69f-4734-ba33-436879200784-kube-api-access-nj995\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.832481 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-config-data-custom\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.832510 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/448fcb03-9b59-4e38-8818-411c7c2dfcf0-logs\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.832568 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzgfr\" (UniqueName: \"kubernetes.io/projected/448fcb03-9b59-4e38-8818-411c7c2dfcf0-kube-api-access-bzgfr\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.832639 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-scripts\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.832684 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-config\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.832714 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-dns-svc\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.833597 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-dns-svc\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.834110 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.834999 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.835545 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.836389 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-config\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.877049 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj995\" (UniqueName: \"kubernetes.io/projected/0f95f00d-d69f-4734-ba33-436879200784-kube-api-access-nj995\") pod \"dnsmasq-dns-5784cf869f-frg9p\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.936231 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzgfr\" (UniqueName: \"kubernetes.io/projected/448fcb03-9b59-4e38-8818-411c7c2dfcf0-kube-api-access-bzgfr\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.936349 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-scripts\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.936401 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/448fcb03-9b59-4e38-8818-411c7c2dfcf0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.936458 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.936477 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-config-data\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.936501 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-config-data-custom\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.936515 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/448fcb03-9b59-4e38-8818-411c7c2dfcf0-logs\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.936993 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/448fcb03-9b59-4e38-8818-411c7c2dfcf0-logs\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.939449 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/448fcb03-9b59-4e38-8818-411c7c2dfcf0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.942811 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.947232 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-scripts\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.948112 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-config-data-custom\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.949843 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-config-data\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:22 crc kubenswrapper[4814]: I0227 16:44:22.959098 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzgfr\" (UniqueName: \"kubernetes.io/projected/448fcb03-9b59-4e38-8818-411c7c2dfcf0-kube-api-access-bzgfr\") pod \"cinder-api-0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " pod="openstack/cinder-api-0" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.010143 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.037010 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.194769 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.345614 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-config-data-custom\") pod \"3584eaa9-9218-4dca-848c-1781f316e4a2\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.345788 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-internal-tls-certs\") pod \"3584eaa9-9218-4dca-848c-1781f316e4a2\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.345817 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-public-tls-certs\") pod \"3584eaa9-9218-4dca-848c-1781f316e4a2\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.345848 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-config-data\") pod \"3584eaa9-9218-4dca-848c-1781f316e4a2\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.346485 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3584eaa9-9218-4dca-848c-1781f316e4a2-logs\") pod \"3584eaa9-9218-4dca-848c-1781f316e4a2\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.346512 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-combined-ca-bundle\") pod \"3584eaa9-9218-4dca-848c-1781f316e4a2\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.346548 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98s7g\" (UniqueName: \"kubernetes.io/projected/3584eaa9-9218-4dca-848c-1781f316e4a2-kube-api-access-98s7g\") pod \"3584eaa9-9218-4dca-848c-1781f316e4a2\" (UID: \"3584eaa9-9218-4dca-848c-1781f316e4a2\") " Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.346887 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3584eaa9-9218-4dca-848c-1781f316e4a2-logs" (OuterVolumeSpecName: "logs") pod "3584eaa9-9218-4dca-848c-1781f316e4a2" (UID: "3584eaa9-9218-4dca-848c-1781f316e4a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.352003 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3584eaa9-9218-4dca-848c-1781f316e4a2-kube-api-access-98s7g" (OuterVolumeSpecName: "kube-api-access-98s7g") pod "3584eaa9-9218-4dca-848c-1781f316e4a2" (UID: "3584eaa9-9218-4dca-848c-1781f316e4a2"). InnerVolumeSpecName "kube-api-access-98s7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.365073 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3584eaa9-9218-4dca-848c-1781f316e4a2" (UID: "3584eaa9-9218-4dca-848c-1781f316e4a2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.408204 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3584eaa9-9218-4dca-848c-1781f316e4a2" (UID: "3584eaa9-9218-4dca-848c-1781f316e4a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.412495 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-config-data" (OuterVolumeSpecName: "config-data") pod "3584eaa9-9218-4dca-848c-1781f316e4a2" (UID: "3584eaa9-9218-4dca-848c-1781f316e4a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.434170 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3584eaa9-9218-4dca-848c-1781f316e4a2" (UID: "3584eaa9-9218-4dca-848c-1781f316e4a2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.448907 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3584eaa9-9218-4dca-848c-1781f316e4a2-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.448937 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.448947 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98s7g\" (UniqueName: \"kubernetes.io/projected/3584eaa9-9218-4dca-848c-1781f316e4a2-kube-api-access-98s7g\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.448959 4814 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.448967 4814 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.448975 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.450345 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3584eaa9-9218-4dca-848c-1781f316e4a2" (UID: "3584eaa9-9218-4dca-848c-1781f316e4a2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.473803 4814 generic.go:334] "Generic (PLEG): container finished" podID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" containerID="146c26d9df94e91bd53b52abca4f6443179c2c8e2365581e20b12612c0129aca" exitCode=0 Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.473833 4814 generic.go:334] "Generic (PLEG): container finished" podID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" containerID="fe992a34a5bd33198fa76c1d351deb8c6a8cb1d10babbc60cf65c741734db762" exitCode=2 Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.473868 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01245f0c-38fe-4f9a-93e5-c5c5d5f32259","Type":"ContainerDied","Data":"146c26d9df94e91bd53b52abca4f6443179c2c8e2365581e20b12612c0129aca"} Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.473894 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01245f0c-38fe-4f9a-93e5-c5c5d5f32259","Type":"ContainerDied","Data":"fe992a34a5bd33198fa76c1d351deb8c6a8cb1d10babbc60cf65c741734db762"} Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.475320 4814 generic.go:334] "Generic (PLEG): container finished" podID="3584eaa9-9218-4dca-848c-1781f316e4a2" containerID="f36a811f0c7b74a3b41ec489863d52d12d465987224d9968a5e7b15e60acf8ec" exitCode=0 Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.475342 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-855bdc9fdb-m79h8" event={"ID":"3584eaa9-9218-4dca-848c-1781f316e4a2","Type":"ContainerDied","Data":"f36a811f0c7b74a3b41ec489863d52d12d465987224d9968a5e7b15e60acf8ec"} Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.475359 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-855bdc9fdb-m79h8" event={"ID":"3584eaa9-9218-4dca-848c-1781f316e4a2","Type":"ContainerDied","Data":"1b2c9e11000a8e15321ebb9234f502702be29180b7977e4a1c31bd2fc8e87b7e"} Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.475374 4814 scope.go:117] "RemoveContainer" containerID="f36a811f0c7b74a3b41ec489863d52d12d465987224d9968a5e7b15e60acf8ec" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.475467 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-855bdc9fdb-m79h8" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.512831 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.514631 4814 scope.go:117] "RemoveContainer" containerID="d2f2e337a2622ede196d96b51299a1cb5497818091db374683707fd677c69485" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.523328 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-855bdc9fdb-m79h8"] Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.531361 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-855bdc9fdb-m79h8"] Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.534828 4814 scope.go:117] "RemoveContainer" containerID="f36a811f0c7b74a3b41ec489863d52d12d465987224d9968a5e7b15e60acf8ec" Feb 27 16:44:23 crc kubenswrapper[4814]: E0227 16:44:23.535388 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f36a811f0c7b74a3b41ec489863d52d12d465987224d9968a5e7b15e60acf8ec\": container with ID starting with f36a811f0c7b74a3b41ec489863d52d12d465987224d9968a5e7b15e60acf8ec not found: ID does not exist" containerID="f36a811f0c7b74a3b41ec489863d52d12d465987224d9968a5e7b15e60acf8ec" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.535431 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f36a811f0c7b74a3b41ec489863d52d12d465987224d9968a5e7b15e60acf8ec"} err="failed to get container status \"f36a811f0c7b74a3b41ec489863d52d12d465987224d9968a5e7b15e60acf8ec\": rpc error: code = NotFound desc = could not find container \"f36a811f0c7b74a3b41ec489863d52d12d465987224d9968a5e7b15e60acf8ec\": container with ID starting with f36a811f0c7b74a3b41ec489863d52d12d465987224d9968a5e7b15e60acf8ec not found: ID does not exist" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.535457 4814 scope.go:117] "RemoveContainer" containerID="d2f2e337a2622ede196d96b51299a1cb5497818091db374683707fd677c69485" Feb 27 16:44:23 crc kubenswrapper[4814]: E0227 16:44:23.535991 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2f2e337a2622ede196d96b51299a1cb5497818091db374683707fd677c69485\": container with ID starting with d2f2e337a2622ede196d96b51299a1cb5497818091db374683707fd677c69485 not found: ID does not exist" containerID="d2f2e337a2622ede196d96b51299a1cb5497818091db374683707fd677c69485" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.536029 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2f2e337a2622ede196d96b51299a1cb5497818091db374683707fd677c69485"} err="failed to get container status \"d2f2e337a2622ede196d96b51299a1cb5497818091db374683707fd677c69485\": rpc error: code = NotFound desc = could not find container \"d2f2e337a2622ede196d96b51299a1cb5497818091db374683707fd677c69485\": container with ID starting with d2f2e337a2622ede196d96b51299a1cb5497818091db374683707fd677c69485 not found: ID does not exist" Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.550543 4814 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3584eaa9-9218-4dca-848c-1781f316e4a2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:23 crc kubenswrapper[4814]: W0227 16:44:23.642287 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f95f00d_d69f_4734_ba33_436879200784.slice/crio-72ab4abb88e1f8871ee0a6289805887ab2cac8acd6c0b2a2e1f46bb4da914ae6 WatchSource:0}: Error finding container 72ab4abb88e1f8871ee0a6289805887ab2cac8acd6c0b2a2e1f46bb4da914ae6: Status 404 returned error can't find the container with id 72ab4abb88e1f8871ee0a6289805887ab2cac8acd6c0b2a2e1f46bb4da914ae6 Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.643687 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-frg9p"] Feb 27 16:44:23 crc kubenswrapper[4814]: I0227 16:44:23.713688 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 27 16:44:24 crc kubenswrapper[4814]: I0227 16:44:24.523749 4814 generic.go:334] "Generic (PLEG): container finished" podID="0f95f00d-d69f-4734-ba33-436879200784" containerID="7180d59d8e49f1f7359649d5407887c77fe44265b021497b2452b85f7866d457" exitCode=0 Feb 27 16:44:24 crc kubenswrapper[4814]: I0227 16:44:24.542335 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3584eaa9-9218-4dca-848c-1781f316e4a2" path="/var/lib/kubelet/pods/3584eaa9-9218-4dca-848c-1781f316e4a2/volumes" Feb 27 16:44:24 crc kubenswrapper[4814]: I0227 16:44:24.543024 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0b8557d-2397-4c7d-b6ab-6044db5f0129" path="/var/lib/kubelet/pods/c0b8557d-2397-4c7d-b6ab-6044db5f0129/volumes" Feb 27 16:44:24 crc kubenswrapper[4814]: I0227 16:44:24.543749 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-frg9p" event={"ID":"0f95f00d-d69f-4734-ba33-436879200784","Type":"ContainerDied","Data":"7180d59d8e49f1f7359649d5407887c77fe44265b021497b2452b85f7866d457"} Feb 27 16:44:24 crc kubenswrapper[4814]: I0227 16:44:24.543786 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-frg9p" event={"ID":"0f95f00d-d69f-4734-ba33-436879200784","Type":"ContainerStarted","Data":"72ab4abb88e1f8871ee0a6289805887ab2cac8acd6c0b2a2e1f46bb4da914ae6"} Feb 27 16:44:24 crc kubenswrapper[4814]: I0227 16:44:24.543804 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d","Type":"ContainerStarted","Data":"f064ca6ac0dabc21e42d893bbf80c09adb43a558f64db6c41616c03d32b4e61f"} Feb 27 16:44:24 crc kubenswrapper[4814]: I0227 16:44:24.543817 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"448fcb03-9b59-4e38-8818-411c7c2dfcf0","Type":"ContainerStarted","Data":"8ea1d4fae8aca57cb4142026a7e89d76148c42d443c58479650a9a7a51e57f4a"} Feb 27 16:44:24 crc kubenswrapper[4814]: I0227 16:44:24.642805 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 27 16:44:25 crc kubenswrapper[4814]: I0227 16:44:25.556374 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d","Type":"ContainerStarted","Data":"09fd568033f6af546f8967f239ffef8a31d2e426cfb9c6829b4021ce8dac1c30"} Feb 27 16:44:25 crc kubenswrapper[4814]: I0227 16:44:25.560385 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"448fcb03-9b59-4e38-8818-411c7c2dfcf0","Type":"ContainerStarted","Data":"0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842"} Feb 27 16:44:25 crc kubenswrapper[4814]: I0227 16:44:25.560428 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"448fcb03-9b59-4e38-8818-411c7c2dfcf0","Type":"ContainerStarted","Data":"ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae"} Feb 27 16:44:25 crc kubenswrapper[4814]: I0227 16:44:25.560557 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="448fcb03-9b59-4e38-8818-411c7c2dfcf0" containerName="cinder-api-log" containerID="cri-o://ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae" gracePeriod=30 Feb 27 16:44:25 crc kubenswrapper[4814]: I0227 16:44:25.560779 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 27 16:44:25 crc kubenswrapper[4814]: I0227 16:44:25.560805 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="448fcb03-9b59-4e38-8818-411c7c2dfcf0" containerName="cinder-api" containerID="cri-o://0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842" gracePeriod=30 Feb 27 16:44:25 crc kubenswrapper[4814]: I0227 16:44:25.581148 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-frg9p" event={"ID":"0f95f00d-d69f-4734-ba33-436879200784","Type":"ContainerStarted","Data":"5dfc7a6c8a51195c471ca7fcf381df65e5441358bcbf95788bec2b01f67f6444"} Feb 27 16:44:25 crc kubenswrapper[4814]: I0227 16:44:25.581375 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:25 crc kubenswrapper[4814]: I0227 16:44:25.614741 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.6147202739999997 podStartE2EDuration="3.614720274s" podCreationTimestamp="2026-02-27 16:44:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:25.577978242 +0000 UTC m=+1278.030603082" watchObservedRunningTime="2026-02-27 16:44:25.614720274 +0000 UTC m=+1278.067345104" Feb 27 16:44:25 crc kubenswrapper[4814]: I0227 16:44:25.619277 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-frg9p" podStartSLOduration=3.619250036 podStartE2EDuration="3.619250036s" podCreationTimestamp="2026-02-27 16:44:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:25.611827054 +0000 UTC m=+1278.064451884" watchObservedRunningTime="2026-02-27 16:44:25.619250036 +0000 UTC m=+1278.071874866" Feb 27 16:44:25 crc kubenswrapper[4814]: I0227 16:44:25.787443 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:44:25 crc kubenswrapper[4814]: I0227 16:44:25.791029 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-5f986b6894-p4tf6" podUID="883b18b4-e86b-4077-8865-ff24ae4838e1" containerName="neutron-httpd" probeResult="failure" output="Get \"http://10.217.0.156:9696/\": dial tcp 10.217.0.156:9696: connect: connection refused" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.456030 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.509666 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-scripts\") pod \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.509725 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/448fcb03-9b59-4e38-8818-411c7c2dfcf0-etc-machine-id\") pod \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.509776 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-config-data-custom\") pod \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.509821 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-combined-ca-bundle\") pod \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.509887 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/448fcb03-9b59-4e38-8818-411c7c2dfcf0-logs\") pod \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.509968 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-config-data\") pod \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.510156 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzgfr\" (UniqueName: \"kubernetes.io/projected/448fcb03-9b59-4e38-8818-411c7c2dfcf0-kube-api-access-bzgfr\") pod \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\" (UID: \"448fcb03-9b59-4e38-8818-411c7c2dfcf0\") " Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.511138 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/448fcb03-9b59-4e38-8818-411c7c2dfcf0-logs" (OuterVolumeSpecName: "logs") pod "448fcb03-9b59-4e38-8818-411c7c2dfcf0" (UID: "448fcb03-9b59-4e38-8818-411c7c2dfcf0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.515169 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/448fcb03-9b59-4e38-8818-411c7c2dfcf0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "448fcb03-9b59-4e38-8818-411c7c2dfcf0" (UID: "448fcb03-9b59-4e38-8818-411c7c2dfcf0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.518607 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/448fcb03-9b59-4e38-8818-411c7c2dfcf0-kube-api-access-bzgfr" (OuterVolumeSpecName: "kube-api-access-bzgfr") pod "448fcb03-9b59-4e38-8818-411c7c2dfcf0" (UID: "448fcb03-9b59-4e38-8818-411c7c2dfcf0"). InnerVolumeSpecName "kube-api-access-bzgfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.519117 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-scripts" (OuterVolumeSpecName: "scripts") pod "448fcb03-9b59-4e38-8818-411c7c2dfcf0" (UID: "448fcb03-9b59-4e38-8818-411c7c2dfcf0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.520451 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "448fcb03-9b59-4e38-8818-411c7c2dfcf0" (UID: "448fcb03-9b59-4e38-8818-411c7c2dfcf0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.544939 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "448fcb03-9b59-4e38-8818-411c7c2dfcf0" (UID: "448fcb03-9b59-4e38-8818-411c7c2dfcf0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.570176 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-config-data" (OuterVolumeSpecName: "config-data") pod "448fcb03-9b59-4e38-8818-411c7c2dfcf0" (UID: "448fcb03-9b59-4e38-8818-411c7c2dfcf0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.592657 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d","Type":"ContainerStarted","Data":"e546075b0a9f6a8272373e147fa224c4e48b40826442f33ed338fe7823a85dee"} Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.594328 4814 generic.go:334] "Generic (PLEG): container finished" podID="448fcb03-9b59-4e38-8818-411c7c2dfcf0" containerID="0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842" exitCode=0 Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.594384 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"448fcb03-9b59-4e38-8818-411c7c2dfcf0","Type":"ContainerDied","Data":"0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842"} Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.594405 4814 generic.go:334] "Generic (PLEG): container finished" podID="448fcb03-9b59-4e38-8818-411c7c2dfcf0" containerID="ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae" exitCode=143 Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.594410 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"448fcb03-9b59-4e38-8818-411c7c2dfcf0","Type":"ContainerDied","Data":"ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae"} Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.594426 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"448fcb03-9b59-4e38-8818-411c7c2dfcf0","Type":"ContainerDied","Data":"8ea1d4fae8aca57cb4142026a7e89d76148c42d443c58479650a9a7a51e57f4a"} Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.594444 4814 scope.go:117] "RemoveContainer" containerID="0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.594371 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.615086 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzgfr\" (UniqueName: \"kubernetes.io/projected/448fcb03-9b59-4e38-8818-411c7c2dfcf0-kube-api-access-bzgfr\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.615283 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.615368 4814 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/448fcb03-9b59-4e38-8818-411c7c2dfcf0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.615446 4814 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.615530 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.615610 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/448fcb03-9b59-4e38-8818-411c7c2dfcf0-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.615690 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/448fcb03-9b59-4e38-8818-411c7c2dfcf0-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.620489 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.8758978539999998 podStartE2EDuration="4.620470913s" podCreationTimestamp="2026-02-27 16:44:22 +0000 UTC" firstStartedPulling="2026-02-27 16:44:23.492632717 +0000 UTC m=+1275.945257547" lastFinishedPulling="2026-02-27 16:44:24.237205776 +0000 UTC m=+1276.689830606" observedRunningTime="2026-02-27 16:44:26.616478259 +0000 UTC m=+1279.069103109" watchObservedRunningTime="2026-02-27 16:44:26.620470913 +0000 UTC m=+1279.073095743" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.691979 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.700128 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.708551 4814 scope.go:117] "RemoveContainer" containerID="ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.718656 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 27 16:44:26 crc kubenswrapper[4814]: E0227 16:44:26.719072 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="448fcb03-9b59-4e38-8818-411c7c2dfcf0" containerName="cinder-api" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.719085 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="448fcb03-9b59-4e38-8818-411c7c2dfcf0" containerName="cinder-api" Feb 27 16:44:26 crc kubenswrapper[4814]: E0227 16:44:26.719099 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3584eaa9-9218-4dca-848c-1781f316e4a2" containerName="barbican-api-log" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.719106 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="3584eaa9-9218-4dca-848c-1781f316e4a2" containerName="barbican-api-log" Feb 27 16:44:26 crc kubenswrapper[4814]: E0227 16:44:26.719119 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="448fcb03-9b59-4e38-8818-411c7c2dfcf0" containerName="cinder-api-log" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.719127 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="448fcb03-9b59-4e38-8818-411c7c2dfcf0" containerName="cinder-api-log" Feb 27 16:44:26 crc kubenswrapper[4814]: E0227 16:44:26.719145 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3584eaa9-9218-4dca-848c-1781f316e4a2" containerName="barbican-api" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.719150 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="3584eaa9-9218-4dca-848c-1781f316e4a2" containerName="barbican-api" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.719403 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="3584eaa9-9218-4dca-848c-1781f316e4a2" containerName="barbican-api" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.719419 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="3584eaa9-9218-4dca-848c-1781f316e4a2" containerName="barbican-api-log" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.719433 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="448fcb03-9b59-4e38-8818-411c7c2dfcf0" containerName="cinder-api-log" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.719444 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="448fcb03-9b59-4e38-8818-411c7c2dfcf0" containerName="cinder-api" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.720334 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.724437 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.724677 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.724787 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.741626 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.759735 4814 scope.go:117] "RemoveContainer" containerID="0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842" Feb 27 16:44:26 crc kubenswrapper[4814]: E0227 16:44:26.760401 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842\": container with ID starting with 0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842 not found: ID does not exist" containerID="0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.760513 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842"} err="failed to get container status \"0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842\": rpc error: code = NotFound desc = could not find container \"0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842\": container with ID starting with 0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842 not found: ID does not exist" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.760701 4814 scope.go:117] "RemoveContainer" containerID="ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae" Feb 27 16:44:26 crc kubenswrapper[4814]: E0227 16:44:26.763914 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae\": container with ID starting with ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae not found: ID does not exist" containerID="ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.763963 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae"} err="failed to get container status \"ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae\": rpc error: code = NotFound desc = could not find container \"ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae\": container with ID starting with ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae not found: ID does not exist" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.763997 4814 scope.go:117] "RemoveContainer" containerID="0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.764353 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842"} err="failed to get container status \"0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842\": rpc error: code = NotFound desc = could not find container \"0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842\": container with ID starting with 0582ec5cb091bf4ffcfb7da7290d3c1d6c69f949874428605ac091eee8a6f842 not found: ID does not exist" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.764375 4814 scope.go:117] "RemoveContainer" containerID="ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.764562 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae"} err="failed to get container status \"ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae\": rpc error: code = NotFound desc = could not find container \"ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae\": container with ID starting with ade245c456cee9c6666cd596ae8e48f3e34351ed3aeaa608f19f35d58422a5ae not found: ID does not exist" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.820169 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-public-tls-certs\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.820225 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-scripts\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.820324 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.820368 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41da1428-a337-48a3-9609-35907e0d4955-logs\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.820536 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41da1428-a337-48a3-9609-35907e0d4955-etc-machine-id\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.820655 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-config-data-custom\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.820704 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-config-data\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.820731 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c54gq\" (UniqueName: \"kubernetes.io/projected/41da1428-a337-48a3-9609-35907e0d4955-kube-api-access-c54gq\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.820770 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.922570 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-public-tls-certs\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.922645 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-scripts\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.922690 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.922718 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41da1428-a337-48a3-9609-35907e0d4955-logs\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.922758 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41da1428-a337-48a3-9609-35907e0d4955-etc-machine-id\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.922814 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-config-data-custom\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.922851 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-config-data\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.922875 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c54gq\" (UniqueName: \"kubernetes.io/projected/41da1428-a337-48a3-9609-35907e0d4955-kube-api-access-c54gq\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.922910 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.923028 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41da1428-a337-48a3-9609-35907e0d4955-etc-machine-id\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.923378 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41da1428-a337-48a3-9609-35907e0d4955-logs\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.928062 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-public-tls-certs\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.928508 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-scripts\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.928980 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.929557 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.930276 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-config-data-custom\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.931000 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41da1428-a337-48a3-9609-35907e0d4955-config-data\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:26 crc kubenswrapper[4814]: I0227 16:44:26.945392 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c54gq\" (UniqueName: \"kubernetes.io/projected/41da1428-a337-48a3-9609-35907e0d4955-kube-api-access-c54gq\") pod \"cinder-api-0\" (UID: \"41da1428-a337-48a3-9609-35907e0d4955\") " pod="openstack/cinder-api-0" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.080983 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.573035 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.602372 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"41da1428-a337-48a3-9609-35907e0d4955","Type":"ContainerStarted","Data":"68a1521b5fb69136b31eeec1ca2c246fb2de68b8152aaa4746a5554068d42421"} Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.606288 4814 generic.go:334] "Generic (PLEG): container finished" podID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" containerID="c52d992512b39f9030e3eb2b20244de9ced5f2493ef47f17365e81f5d87b6cf2" exitCode=0 Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.607050 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01245f0c-38fe-4f9a-93e5-c5c5d5f32259","Type":"ContainerDied","Data":"c52d992512b39f9030e3eb2b20244de9ced5f2493ef47f17365e81f5d87b6cf2"} Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.607079 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01245f0c-38fe-4f9a-93e5-c5c5d5f32259","Type":"ContainerDied","Data":"f302b88507b3e57c745e98835c09044d94519da0717d49898477cae134d2a018"} Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.607094 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f302b88507b3e57c745e98835c09044d94519da0717d49898477cae134d2a018" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.661366 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.738126 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-sg-core-conf-yaml\") pod \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.738282 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-run-httpd\") pod \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.738411 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-log-httpd\") pod \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.738482 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-scripts\") pod \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.738515 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-config-data\") pod \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.738600 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdqh8\" (UniqueName: \"kubernetes.io/projected/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-kube-api-access-fdqh8\") pod \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.739317 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-combined-ca-bundle\") pod \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\" (UID: \"01245f0c-38fe-4f9a-93e5-c5c5d5f32259\") " Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.738974 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "01245f0c-38fe-4f9a-93e5-c5c5d5f32259" (UID: "01245f0c-38fe-4f9a-93e5-c5c5d5f32259"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.739151 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "01245f0c-38fe-4f9a-93e5-c5c5d5f32259" (UID: "01245f0c-38fe-4f9a-93e5-c5c5d5f32259"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.739904 4814 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.739927 4814 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.745329 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-kube-api-access-fdqh8" (OuterVolumeSpecName: "kube-api-access-fdqh8") pod "01245f0c-38fe-4f9a-93e5-c5c5d5f32259" (UID: "01245f0c-38fe-4f9a-93e5-c5c5d5f32259"). InnerVolumeSpecName "kube-api-access-fdqh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.745565 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-scripts" (OuterVolumeSpecName: "scripts") pod "01245f0c-38fe-4f9a-93e5-c5c5d5f32259" (UID: "01245f0c-38fe-4f9a-93e5-c5c5d5f32259"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.759752 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.767058 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "01245f0c-38fe-4f9a-93e5-c5c5d5f32259" (UID: "01245f0c-38fe-4f9a-93e5-c5c5d5f32259"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.790830 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01245f0c-38fe-4f9a-93e5-c5c5d5f32259" (UID: "01245f0c-38fe-4f9a-93e5-c5c5d5f32259"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.821392 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-config-data" (OuterVolumeSpecName: "config-data") pod "01245f0c-38fe-4f9a-93e5-c5c5d5f32259" (UID: "01245f0c-38fe-4f9a-93e5-c5c5d5f32259"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.841243 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdqh8\" (UniqueName: \"kubernetes.io/projected/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-kube-api-access-fdqh8\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.841296 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.841309 4814 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.841323 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.841334 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01245f0c-38fe-4f9a-93e5-c5c5d5f32259-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:27 crc kubenswrapper[4814]: I0227 16:44:27.991654 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.499372 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="448fcb03-9b59-4e38-8818-411c7c2dfcf0" path="/var/lib/kubelet/pods/448fcb03-9b59-4e38-8818-411c7c2dfcf0/volumes" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.618306 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"41da1428-a337-48a3-9609-35907e0d4955","Type":"ContainerStarted","Data":"9d33c1353e1a68968338c2fd5fbe938436fcb2ecd3c6ccbdd4a00856f0debd9a"} Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.618396 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.677336 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.715419 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.729773 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:44:28 crc kubenswrapper[4814]: E0227 16:44:28.730471 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" containerName="sg-core" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.730489 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" containerName="sg-core" Feb 27 16:44:28 crc kubenswrapper[4814]: E0227 16:44:28.730514 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" containerName="ceilometer-notification-agent" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.730521 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" containerName="ceilometer-notification-agent" Feb 27 16:44:28 crc kubenswrapper[4814]: E0227 16:44:28.730536 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" containerName="proxy-httpd" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.730544 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" containerName="proxy-httpd" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.730814 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" containerName="proxy-httpd" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.730845 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" containerName="sg-core" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.730862 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" containerName="ceilometer-notification-agent" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.738565 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.741393 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.741690 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.741781 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.873030 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ad0ff7ef-a759-4820-8c49-d971adb89ff7-log-httpd\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.873080 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fcnv\" (UniqueName: \"kubernetes.io/projected/ad0ff7ef-a759-4820-8c49-d971adb89ff7-kube-api-access-4fcnv\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.873122 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.873199 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.873274 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-scripts\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.873311 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-config-data\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.873327 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ad0ff7ef-a759-4820-8c49-d971adb89ff7-run-httpd\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.975049 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ad0ff7ef-a759-4820-8c49-d971adb89ff7-log-httpd\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.975138 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fcnv\" (UniqueName: \"kubernetes.io/projected/ad0ff7ef-a759-4820-8c49-d971adb89ff7-kube-api-access-4fcnv\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.975223 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.975280 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.975381 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-scripts\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.975456 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-config-data\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.975487 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ad0ff7ef-a759-4820-8c49-d971adb89ff7-run-httpd\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.975946 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ad0ff7ef-a759-4820-8c49-d971adb89ff7-log-httpd\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.976097 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ad0ff7ef-a759-4820-8c49-d971adb89ff7-run-httpd\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.981853 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-scripts\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.988330 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.989142 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:28 crc kubenswrapper[4814]: I0227 16:44:28.995062 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-config-data\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:29 crc kubenswrapper[4814]: I0227 16:44:29.007748 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fcnv\" (UniqueName: \"kubernetes.io/projected/ad0ff7ef-a759-4820-8c49-d971adb89ff7-kube-api-access-4fcnv\") pod \"ceilometer-0\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " pod="openstack/ceilometer-0" Feb 27 16:44:29 crc kubenswrapper[4814]: I0227 16:44:29.057485 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:44:29 crc kubenswrapper[4814]: I0227 16:44:29.602326 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:44:29 crc kubenswrapper[4814]: W0227 16:44:29.616868 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad0ff7ef_a759_4820_8c49_d971adb89ff7.slice/crio-3134271f6f40fb17bfc70e878327952f1e2b51c8007c37d1b5e29524d3d37e80 WatchSource:0}: Error finding container 3134271f6f40fb17bfc70e878327952f1e2b51c8007c37d1b5e29524d3d37e80: Status 404 returned error can't find the container with id 3134271f6f40fb17bfc70e878327952f1e2b51c8007c37d1b5e29524d3d37e80 Feb 27 16:44:29 crc kubenswrapper[4814]: I0227 16:44:29.634660 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ad0ff7ef-a759-4820-8c49-d971adb89ff7","Type":"ContainerStarted","Data":"3134271f6f40fb17bfc70e878327952f1e2b51c8007c37d1b5e29524d3d37e80"} Feb 27 16:44:29 crc kubenswrapper[4814]: I0227 16:44:29.638084 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"41da1428-a337-48a3-9609-35907e0d4955","Type":"ContainerStarted","Data":"db66c1d1d852035aebf9bfe846d6d483704c57335513ce251ff66ecbdbbafe13"} Feb 27 16:44:29 crc kubenswrapper[4814]: I0227 16:44:29.638986 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 27 16:44:29 crc kubenswrapper[4814]: I0227 16:44:29.661465 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.661451727 podStartE2EDuration="3.661451727s" podCreationTimestamp="2026-02-27 16:44:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:29.658564787 +0000 UTC m=+1282.111189617" watchObservedRunningTime="2026-02-27 16:44:29.661451727 +0000 UTC m=+1282.114076557" Feb 27 16:44:30 crc kubenswrapper[4814]: I0227 16:44:30.498130 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01245f0c-38fe-4f9a-93e5-c5c5d5f32259" path="/var/lib/kubelet/pods/01245f0c-38fe-4f9a-93e5-c5c5d5f32259/volumes" Feb 27 16:44:30 crc kubenswrapper[4814]: I0227 16:44:30.650244 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ad0ff7ef-a759-4820-8c49-d971adb89ff7","Type":"ContainerStarted","Data":"99163a979e91bbcbfa552f4f52c416a375c6848873571d31e1d6b144edb2089d"} Feb 27 16:44:31 crc kubenswrapper[4814]: I0227 16:44:31.446047 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:44:31 crc kubenswrapper[4814]: I0227 16:44:31.638815 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:44:31 crc kubenswrapper[4814]: I0227 16:44:31.665942 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ad0ff7ef-a759-4820-8c49-d971adb89ff7","Type":"ContainerStarted","Data":"d44aff35cb251c05ab91e5f7b2b03c03c4d744f9c8554efc9d2a35a3f56e9e5f"} Feb 27 16:44:32 crc kubenswrapper[4814]: I0227 16:44:32.985887 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.011982 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.065985 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.127297 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-2s2xt"] Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.129091 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" podUID="097c7fc7-f42b-494c-8d25-f7e76a8350b5" containerName="dnsmasq-dns" containerID="cri-o://aeee546fd66f4c8b06a06463a4c7549770e19ff37831945befcd27b375a8a0f8" gracePeriod=10 Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.168586 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.681381 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-65784f76f6-f2pcp" Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.687223 4814 generic.go:334] "Generic (PLEG): container finished" podID="097c7fc7-f42b-494c-8d25-f7e76a8350b5" containerID="aeee546fd66f4c8b06a06463a4c7549770e19ff37831945befcd27b375a8a0f8" exitCode=0 Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.687286 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" event={"ID":"097c7fc7-f42b-494c-8d25-f7e76a8350b5","Type":"ContainerDied","Data":"aeee546fd66f4c8b06a06463a4c7549770e19ff37831945befcd27b375a8a0f8"} Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.687438 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" containerName="cinder-scheduler" containerID="cri-o://09fd568033f6af546f8967f239ffef8a31d2e426cfb9c6829b4021ce8dac1c30" gracePeriod=30 Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.687438 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" containerName="probe" containerID="cri-o://e546075b0a9f6a8272373e147fa224c4e48b40826442f33ed338fe7823a85dee" gracePeriod=30 Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.758934 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-845d8f46cb-qhstj"] Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.759975 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-845d8f46cb-qhstj" podUID="6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" containerName="horizon-log" containerID="cri-o://19807c67c77d45f3562d439c93d9fa848a39b696dec9481c4ec2a11a35968141" gracePeriod=30 Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.760402 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-845d8f46cb-qhstj" podUID="6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" containerName="horizon" containerID="cri-o://5c624a9feb184d9603fa9c3597fae610b34d3632ba8c53b56374223a4f027fd2" gracePeriod=30 Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.828982 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.899229 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-dns-swift-storage-0\") pod \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.899300 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-ovsdbserver-nb\") pod \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.899385 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-config\") pod \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.899431 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccrlt\" (UniqueName: \"kubernetes.io/projected/097c7fc7-f42b-494c-8d25-f7e76a8350b5-kube-api-access-ccrlt\") pod \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.899491 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-dns-svc\") pod \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.899574 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-ovsdbserver-sb\") pod \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\" (UID: \"097c7fc7-f42b-494c-8d25-f7e76a8350b5\") " Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.907441 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/097c7fc7-f42b-494c-8d25-f7e76a8350b5-kube-api-access-ccrlt" (OuterVolumeSpecName: "kube-api-access-ccrlt") pod "097c7fc7-f42b-494c-8d25-f7e76a8350b5" (UID: "097c7fc7-f42b-494c-8d25-f7e76a8350b5"). InnerVolumeSpecName "kube-api-access-ccrlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.981723 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-config" (OuterVolumeSpecName: "config") pod "097c7fc7-f42b-494c-8d25-f7e76a8350b5" (UID: "097c7fc7-f42b-494c-8d25-f7e76a8350b5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.981734 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "097c7fc7-f42b-494c-8d25-f7e76a8350b5" (UID: "097c7fc7-f42b-494c-8d25-f7e76a8350b5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:33 crc kubenswrapper[4814]: I0227 16:44:33.991965 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "097c7fc7-f42b-494c-8d25-f7e76a8350b5" (UID: "097c7fc7-f42b-494c-8d25-f7e76a8350b5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.001571 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.001599 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.001608 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccrlt\" (UniqueName: \"kubernetes.io/projected/097c7fc7-f42b-494c-8d25-f7e76a8350b5-kube-api-access-ccrlt\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.001618 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.003024 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "097c7fc7-f42b-494c-8d25-f7e76a8350b5" (UID: "097c7fc7-f42b-494c-8d25-f7e76a8350b5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.011982 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "097c7fc7-f42b-494c-8d25-f7e76a8350b5" (UID: "097c7fc7-f42b-494c-8d25-f7e76a8350b5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.103893 4814 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.104185 4814 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/097c7fc7-f42b-494c-8d25-f7e76a8350b5-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.148762 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-69895c5d4f-4bpj5" Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.227884 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-85f77557cc-k22bs"] Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.230524 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-85f77557cc-k22bs" podUID="131dd92c-e25a-4385-9380-84e286f09a43" containerName="neutron-httpd" containerID="cri-o://bdaf42af64df4a5bbcec9295585e7515e63e2138955c345374e7af4e4023a5e8" gracePeriod=30 Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.228409 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-85f77557cc-k22bs" podUID="131dd92c-e25a-4385-9380-84e286f09a43" containerName="neutron-api" containerID="cri-o://d13649093c51431f472803e631c21e26b1fdd0d8c979de1fc946481976409c01" gracePeriod=30 Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.554859 4814 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod3efc1a92-93e1-45a5-86c6-263ad2d4c36b"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod3efc1a92-93e1-45a5-86c6-263ad2d4c36b] : Timed out while waiting for systemd to remove kubepods-besteffort-pod3efc1a92_93e1_45a5_86c6_263ad2d4c36b.slice" Feb 27 16:44:34 crc kubenswrapper[4814]: E0227 16:44:34.555216 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod3efc1a92-93e1-45a5-86c6-263ad2d4c36b] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod3efc1a92-93e1-45a5-86c6-263ad2d4c36b] : Timed out while waiting for systemd to remove kubepods-besteffort-pod3efc1a92_93e1_45a5_86c6_263ad2d4c36b.slice" pod="openstack/dnsmasq-dns-84b966f6c9-hx978" podUID="3efc1a92-93e1-45a5-86c6-263ad2d4c36b" Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.696649 4814 generic.go:334] "Generic (PLEG): container finished" podID="131dd92c-e25a-4385-9380-84e286f09a43" containerID="bdaf42af64df4a5bbcec9295585e7515e63e2138955c345374e7af4e4023a5e8" exitCode=0 Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.696712 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85f77557cc-k22bs" event={"ID":"131dd92c-e25a-4385-9380-84e286f09a43","Type":"ContainerDied","Data":"bdaf42af64df4a5bbcec9295585e7515e63e2138955c345374e7af4e4023a5e8"} Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.698725 4814 generic.go:334] "Generic (PLEG): container finished" podID="20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" containerID="e546075b0a9f6a8272373e147fa224c4e48b40826442f33ed338fe7823a85dee" exitCode=0 Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.698763 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d","Type":"ContainerDied","Data":"e546075b0a9f6a8272373e147fa224c4e48b40826442f33ed338fe7823a85dee"} Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.700337 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-hx978" Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.700390 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" event={"ID":"097c7fc7-f42b-494c-8d25-f7e76a8350b5","Type":"ContainerDied","Data":"21ede465b3046ed6554972811baa18a99ff32a95bc2656e6a3a249ab80a89970"} Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.700444 4814 scope.go:117] "RemoveContainer" containerID="aeee546fd66f4c8b06a06463a4c7549770e19ff37831945befcd27b375a8a0f8" Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.700962 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-2s2xt" Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.723571 4814 scope.go:117] "RemoveContainer" containerID="3e12afe869bb6c7cd7c682a61ea3c6f27c1f7af9fce87a9707cffd7869f3b249" Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.764059 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-hx978"] Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.784392 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-hx978"] Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.809008 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-2s2xt"] Feb 27 16:44:34 crc kubenswrapper[4814]: I0227 16:44:34.815790 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-2s2xt"] Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.382969 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7877c8f6d6-q56qs" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.435905 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5f986b6894-p4tf6_883b18b4-e86b-4077-8865-ff24ae4838e1/neutron-api/0.log" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.435963 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.540307 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-ovndb-tls-certs\") pod \"883b18b4-e86b-4077-8865-ff24ae4838e1\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.540638 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-httpd-config\") pod \"883b18b4-e86b-4077-8865-ff24ae4838e1\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.540699 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-config\") pod \"883b18b4-e86b-4077-8865-ff24ae4838e1\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.540724 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tj5x6\" (UniqueName: \"kubernetes.io/projected/883b18b4-e86b-4077-8865-ff24ae4838e1-kube-api-access-tj5x6\") pod \"883b18b4-e86b-4077-8865-ff24ae4838e1\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.540759 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-combined-ca-bundle\") pod \"883b18b4-e86b-4077-8865-ff24ae4838e1\" (UID: \"883b18b4-e86b-4077-8865-ff24ae4838e1\") " Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.550555 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/883b18b4-e86b-4077-8865-ff24ae4838e1-kube-api-access-tj5x6" (OuterVolumeSpecName: "kube-api-access-tj5x6") pod "883b18b4-e86b-4077-8865-ff24ae4838e1" (UID: "883b18b4-e86b-4077-8865-ff24ae4838e1"). InnerVolumeSpecName "kube-api-access-tj5x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.551585 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "883b18b4-e86b-4077-8865-ff24ae4838e1" (UID: "883b18b4-e86b-4077-8865-ff24ae4838e1"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.597918 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 27 16:44:35 crc kubenswrapper[4814]: E0227 16:44:35.598301 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="883b18b4-e86b-4077-8865-ff24ae4838e1" containerName="neutron-api" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.598317 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="883b18b4-e86b-4077-8865-ff24ae4838e1" containerName="neutron-api" Feb 27 16:44:35 crc kubenswrapper[4814]: E0227 16:44:35.598328 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="097c7fc7-f42b-494c-8d25-f7e76a8350b5" containerName="init" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.598334 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="097c7fc7-f42b-494c-8d25-f7e76a8350b5" containerName="init" Feb 27 16:44:35 crc kubenswrapper[4814]: E0227 16:44:35.598355 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="883b18b4-e86b-4077-8865-ff24ae4838e1" containerName="neutron-httpd" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.598362 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="883b18b4-e86b-4077-8865-ff24ae4838e1" containerName="neutron-httpd" Feb 27 16:44:35 crc kubenswrapper[4814]: E0227 16:44:35.598383 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="097c7fc7-f42b-494c-8d25-f7e76a8350b5" containerName="dnsmasq-dns" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.598388 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="097c7fc7-f42b-494c-8d25-f7e76a8350b5" containerName="dnsmasq-dns" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.598563 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="883b18b4-e86b-4077-8865-ff24ae4838e1" containerName="neutron-api" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.598583 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="883b18b4-e86b-4077-8865-ff24ae4838e1" containerName="neutron-httpd" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.598608 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="097c7fc7-f42b-494c-8d25-f7e76a8350b5" containerName="dnsmasq-dns" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.599140 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.607672 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.607896 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-rs84b" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.608400 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.612708 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.644266 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cc735347-66c8-4f38-89e9-31345e59ffee-openstack-config-secret\") pod \"openstackclient\" (UID: \"cc735347-66c8-4f38-89e9-31345e59ffee\") " pod="openstack/openstackclient" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.644320 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc735347-66c8-4f38-89e9-31345e59ffee-combined-ca-bundle\") pod \"openstackclient\" (UID: \"cc735347-66c8-4f38-89e9-31345e59ffee\") " pod="openstack/openstackclient" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.644371 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mwlf\" (UniqueName: \"kubernetes.io/projected/cc735347-66c8-4f38-89e9-31345e59ffee-kube-api-access-9mwlf\") pod \"openstackclient\" (UID: \"cc735347-66c8-4f38-89e9-31345e59ffee\") " pod="openstack/openstackclient" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.644722 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cc735347-66c8-4f38-89e9-31345e59ffee-openstack-config\") pod \"openstackclient\" (UID: \"cc735347-66c8-4f38-89e9-31345e59ffee\") " pod="openstack/openstackclient" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.644875 4814 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.644892 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tj5x6\" (UniqueName: \"kubernetes.io/projected/883b18b4-e86b-4077-8865-ff24ae4838e1-kube-api-access-tj5x6\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.649316 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "883b18b4-e86b-4077-8865-ff24ae4838e1" (UID: "883b18b4-e86b-4077-8865-ff24ae4838e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.681333 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-config" (OuterVolumeSpecName: "config") pod "883b18b4-e86b-4077-8865-ff24ae4838e1" (UID: "883b18b4-e86b-4077-8865-ff24ae4838e1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.694941 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "883b18b4-e86b-4077-8865-ff24ae4838e1" (UID: "883b18b4-e86b-4077-8865-ff24ae4838e1"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.717665 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5f986b6894-p4tf6_883b18b4-e86b-4077-8865-ff24ae4838e1/neutron-api/0.log" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.717709 4814 generic.go:334] "Generic (PLEG): container finished" podID="883b18b4-e86b-4077-8865-ff24ae4838e1" containerID="cfa61f32aebd404721137338d8249be5279fd83f1dab45436e53bbca1961e27c" exitCode=137 Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.717735 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f986b6894-p4tf6" event={"ID":"883b18b4-e86b-4077-8865-ff24ae4838e1","Type":"ContainerDied","Data":"cfa61f32aebd404721137338d8249be5279fd83f1dab45436e53bbca1961e27c"} Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.717758 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f986b6894-p4tf6" event={"ID":"883b18b4-e86b-4077-8865-ff24ae4838e1","Type":"ContainerDied","Data":"c05911f14d7f1b2fa25760f3ec14493754772f5170d78ea23abdd7c1a6e9e453"} Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.717775 4814 scope.go:117] "RemoveContainer" containerID="993646e01cc33127d18c148d44f285e775bc2a6765cd92bfeaf24782ce055cf5" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.717858 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f986b6894-p4tf6" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.740295 4814 scope.go:117] "RemoveContainer" containerID="cfa61f32aebd404721137338d8249be5279fd83f1dab45436e53bbca1961e27c" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.746289 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cc735347-66c8-4f38-89e9-31345e59ffee-openstack-config\") pod \"openstackclient\" (UID: \"cc735347-66c8-4f38-89e9-31345e59ffee\") " pod="openstack/openstackclient" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.746369 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cc735347-66c8-4f38-89e9-31345e59ffee-openstack-config-secret\") pod \"openstackclient\" (UID: \"cc735347-66c8-4f38-89e9-31345e59ffee\") " pod="openstack/openstackclient" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.746406 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc735347-66c8-4f38-89e9-31345e59ffee-combined-ca-bundle\") pod \"openstackclient\" (UID: \"cc735347-66c8-4f38-89e9-31345e59ffee\") " pod="openstack/openstackclient" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.746465 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mwlf\" (UniqueName: \"kubernetes.io/projected/cc735347-66c8-4f38-89e9-31345e59ffee-kube-api-access-9mwlf\") pod \"openstackclient\" (UID: \"cc735347-66c8-4f38-89e9-31345e59ffee\") " pod="openstack/openstackclient" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.746534 4814 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.746546 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.746592 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/883b18b4-e86b-4077-8865-ff24ae4838e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.747149 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cc735347-66c8-4f38-89e9-31345e59ffee-openstack-config\") pod \"openstackclient\" (UID: \"cc735347-66c8-4f38-89e9-31345e59ffee\") " pod="openstack/openstackclient" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.749762 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc735347-66c8-4f38-89e9-31345e59ffee-combined-ca-bundle\") pod \"openstackclient\" (UID: \"cc735347-66c8-4f38-89e9-31345e59ffee\") " pod="openstack/openstackclient" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.749824 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cc735347-66c8-4f38-89e9-31345e59ffee-openstack-config-secret\") pod \"openstackclient\" (UID: \"cc735347-66c8-4f38-89e9-31345e59ffee\") " pod="openstack/openstackclient" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.760825 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mwlf\" (UniqueName: \"kubernetes.io/projected/cc735347-66c8-4f38-89e9-31345e59ffee-kube-api-access-9mwlf\") pod \"openstackclient\" (UID: \"cc735347-66c8-4f38-89e9-31345e59ffee\") " pod="openstack/openstackclient" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.815295 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5f986b6894-p4tf6"] Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.823161 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5f986b6894-p4tf6"] Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.825877 4814 scope.go:117] "RemoveContainer" containerID="993646e01cc33127d18c148d44f285e775bc2a6765cd92bfeaf24782ce055cf5" Feb 27 16:44:35 crc kubenswrapper[4814]: E0227 16:44:35.826426 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"993646e01cc33127d18c148d44f285e775bc2a6765cd92bfeaf24782ce055cf5\": container with ID starting with 993646e01cc33127d18c148d44f285e775bc2a6765cd92bfeaf24782ce055cf5 not found: ID does not exist" containerID="993646e01cc33127d18c148d44f285e775bc2a6765cd92bfeaf24782ce055cf5" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.826481 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"993646e01cc33127d18c148d44f285e775bc2a6765cd92bfeaf24782ce055cf5"} err="failed to get container status \"993646e01cc33127d18c148d44f285e775bc2a6765cd92bfeaf24782ce055cf5\": rpc error: code = NotFound desc = could not find container \"993646e01cc33127d18c148d44f285e775bc2a6765cd92bfeaf24782ce055cf5\": container with ID starting with 993646e01cc33127d18c148d44f285e775bc2a6765cd92bfeaf24782ce055cf5 not found: ID does not exist" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.826519 4814 scope.go:117] "RemoveContainer" containerID="cfa61f32aebd404721137338d8249be5279fd83f1dab45436e53bbca1961e27c" Feb 27 16:44:35 crc kubenswrapper[4814]: E0227 16:44:35.827059 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfa61f32aebd404721137338d8249be5279fd83f1dab45436e53bbca1961e27c\": container with ID starting with cfa61f32aebd404721137338d8249be5279fd83f1dab45436e53bbca1961e27c not found: ID does not exist" containerID="cfa61f32aebd404721137338d8249be5279fd83f1dab45436e53bbca1961e27c" Feb 27 16:44:35 crc kubenswrapper[4814]: I0227 16:44:35.827119 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfa61f32aebd404721137338d8249be5279fd83f1dab45436e53bbca1961e27c"} err="failed to get container status \"cfa61f32aebd404721137338d8249be5279fd83f1dab45436e53bbca1961e27c\": rpc error: code = NotFound desc = could not find container \"cfa61f32aebd404721137338d8249be5279fd83f1dab45436e53bbca1961e27c\": container with ID starting with cfa61f32aebd404721137338d8249be5279fd83f1dab45436e53bbca1961e27c not found: ID does not exist" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.017438 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.290345 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.357917 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-combined-ca-bundle\") pod \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.358046 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlfrc\" (UniqueName: \"kubernetes.io/projected/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-kube-api-access-qlfrc\") pod \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.358067 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-etc-machine-id\") pod \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.358197 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-config-data\") pod \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.358223 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-scripts\") pod \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.358238 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-config-data-custom\") pod \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\" (UID: \"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d\") " Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.361777 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" (UID: "20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.362864 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" (UID: "20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.363517 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-kube-api-access-qlfrc" (OuterVolumeSpecName: "kube-api-access-qlfrc") pod "20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" (UID: "20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d"). InnerVolumeSpecName "kube-api-access-qlfrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.384416 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-scripts" (OuterVolumeSpecName: "scripts") pod "20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" (UID: "20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.460400 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" (UID: "20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.461029 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.461101 4814 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.461156 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.461207 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlfrc\" (UniqueName: \"kubernetes.io/projected/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-kube-api-access-qlfrc\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.461303 4814 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.463497 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.502178 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="097c7fc7-f42b-494c-8d25-f7e76a8350b5" path="/var/lib/kubelet/pods/097c7fc7-f42b-494c-8d25-f7e76a8350b5/volumes" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.503408 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3efc1a92-93e1-45a5-86c6-263ad2d4c36b" path="/var/lib/kubelet/pods/3efc1a92-93e1-45a5-86c6-263ad2d4c36b/volumes" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.503864 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="883b18b4-e86b-4077-8865-ff24ae4838e1" path="/var/lib/kubelet/pods/883b18b4-e86b-4077-8865-ff24ae4838e1/volumes" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.526981 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-config-data" (OuterVolumeSpecName: "config-data") pod "20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" (UID: "20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.565109 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.727024 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ad0ff7ef-a759-4820-8c49-d971adb89ff7","Type":"ContainerStarted","Data":"7f7b486d5cf6ef7eb75ff760e556f493eb909730b66a48113f081b82ed9e2d23"} Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.728270 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"cc735347-66c8-4f38-89e9-31345e59ffee","Type":"ContainerStarted","Data":"32c653874563258a6a1971afebe1fbe5b9b60f2be5205c31e9c2c93d4ecff362"} Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.731949 4814 generic.go:334] "Generic (PLEG): container finished" podID="20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" containerID="09fd568033f6af546f8967f239ffef8a31d2e426cfb9c6829b4021ce8dac1c30" exitCode=0 Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.731980 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d","Type":"ContainerDied","Data":"09fd568033f6af546f8967f239ffef8a31d2e426cfb9c6829b4021ce8dac1c30"} Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.732016 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d","Type":"ContainerDied","Data":"f064ca6ac0dabc21e42d893bbf80c09adb43a558f64db6c41616c03d32b4e61f"} Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.732035 4814 scope.go:117] "RemoveContainer" containerID="e546075b0a9f6a8272373e147fa224c4e48b40826442f33ed338fe7823a85dee" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.732043 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.764700 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.766798 4814 scope.go:117] "RemoveContainer" containerID="09fd568033f6af546f8967f239ffef8a31d2e426cfb9c6829b4021ce8dac1c30" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.777215 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.794002 4814 scope.go:117] "RemoveContainer" containerID="e546075b0a9f6a8272373e147fa224c4e48b40826442f33ed338fe7823a85dee" Feb 27 16:44:36 crc kubenswrapper[4814]: E0227 16:44:36.794767 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e546075b0a9f6a8272373e147fa224c4e48b40826442f33ed338fe7823a85dee\": container with ID starting with e546075b0a9f6a8272373e147fa224c4e48b40826442f33ed338fe7823a85dee not found: ID does not exist" containerID="e546075b0a9f6a8272373e147fa224c4e48b40826442f33ed338fe7823a85dee" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.794799 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e546075b0a9f6a8272373e147fa224c4e48b40826442f33ed338fe7823a85dee"} err="failed to get container status \"e546075b0a9f6a8272373e147fa224c4e48b40826442f33ed338fe7823a85dee\": rpc error: code = NotFound desc = could not find container \"e546075b0a9f6a8272373e147fa224c4e48b40826442f33ed338fe7823a85dee\": container with ID starting with e546075b0a9f6a8272373e147fa224c4e48b40826442f33ed338fe7823a85dee not found: ID does not exist" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.794817 4814 scope.go:117] "RemoveContainer" containerID="09fd568033f6af546f8967f239ffef8a31d2e426cfb9c6829b4021ce8dac1c30" Feb 27 16:44:36 crc kubenswrapper[4814]: E0227 16:44:36.795296 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09fd568033f6af546f8967f239ffef8a31d2e426cfb9c6829b4021ce8dac1c30\": container with ID starting with 09fd568033f6af546f8967f239ffef8a31d2e426cfb9c6829b4021ce8dac1c30 not found: ID does not exist" containerID="09fd568033f6af546f8967f239ffef8a31d2e426cfb9c6829b4021ce8dac1c30" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.795352 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09fd568033f6af546f8967f239ffef8a31d2e426cfb9c6829b4021ce8dac1c30"} err="failed to get container status \"09fd568033f6af546f8967f239ffef8a31d2e426cfb9c6829b4021ce8dac1c30\": rpc error: code = NotFound desc = could not find container \"09fd568033f6af546f8967f239ffef8a31d2e426cfb9c6829b4021ce8dac1c30\": container with ID starting with 09fd568033f6af546f8967f239ffef8a31d2e426cfb9c6829b4021ce8dac1c30 not found: ID does not exist" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.805047 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 16:44:36 crc kubenswrapper[4814]: E0227 16:44:36.805505 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" containerName="probe" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.805526 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" containerName="probe" Feb 27 16:44:36 crc kubenswrapper[4814]: E0227 16:44:36.805558 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" containerName="cinder-scheduler" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.805565 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" containerName="cinder-scheduler" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.805720 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" containerName="probe" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.805747 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" containerName="cinder-scheduler" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.806792 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.809373 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.818501 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.873777 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd1ed9cd-1472-46c3-852f-a41a2a316e40-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.873865 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd1ed9cd-1472-46c3-852f-a41a2a316e40-scripts\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.873906 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjrkp\" (UniqueName: \"kubernetes.io/projected/bd1ed9cd-1472-46c3-852f-a41a2a316e40-kube-api-access-sjrkp\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.873973 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd1ed9cd-1472-46c3-852f-a41a2a316e40-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.874131 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd1ed9cd-1472-46c3-852f-a41a2a316e40-config-data\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.874148 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd1ed9cd-1472-46c3-852f-a41a2a316e40-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.975809 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd1ed9cd-1472-46c3-852f-a41a2a316e40-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.975935 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd1ed9cd-1472-46c3-852f-a41a2a316e40-config-data\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.975954 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd1ed9cd-1472-46c3-852f-a41a2a316e40-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.975974 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd1ed9cd-1472-46c3-852f-a41a2a316e40-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.976007 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd1ed9cd-1472-46c3-852f-a41a2a316e40-scripts\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.976024 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjrkp\" (UniqueName: \"kubernetes.io/projected/bd1ed9cd-1472-46c3-852f-a41a2a316e40-kube-api-access-sjrkp\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.977076 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd1ed9cd-1472-46c3-852f-a41a2a316e40-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.980988 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd1ed9cd-1472-46c3-852f-a41a2a316e40-scripts\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.981226 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd1ed9cd-1472-46c3-852f-a41a2a316e40-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.981223 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd1ed9cd-1472-46c3-852f-a41a2a316e40-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:36 crc kubenswrapper[4814]: I0227 16:44:36.982445 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd1ed9cd-1472-46c3-852f-a41a2a316e40-config-data\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:37 crc kubenswrapper[4814]: I0227 16:44:37.001676 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjrkp\" (UniqueName: \"kubernetes.io/projected/bd1ed9cd-1472-46c3-852f-a41a2a316e40-kube-api-access-sjrkp\") pod \"cinder-scheduler-0\" (UID: \"bd1ed9cd-1472-46c3-852f-a41a2a316e40\") " pod="openstack/cinder-scheduler-0" Feb 27 16:44:37 crc kubenswrapper[4814]: I0227 16:44:37.127685 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 16:44:37 crc kubenswrapper[4814]: I0227 16:44:37.687028 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 16:44:37 crc kubenswrapper[4814]: I0227 16:44:37.753488 4814 generic.go:334] "Generic (PLEG): container finished" podID="6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" containerID="5c624a9feb184d9603fa9c3597fae610b34d3632ba8c53b56374223a4f027fd2" exitCode=0 Feb 27 16:44:37 crc kubenswrapper[4814]: I0227 16:44:37.753550 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-845d8f46cb-qhstj" event={"ID":"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1","Type":"ContainerDied","Data":"5c624a9feb184d9603fa9c3597fae610b34d3632ba8c53b56374223a4f027fd2"} Feb 27 16:44:37 crc kubenswrapper[4814]: I0227 16:44:37.754823 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bd1ed9cd-1472-46c3-852f-a41a2a316e40","Type":"ContainerStarted","Data":"ae55f16e1cbc301f88b6c057059b933d7e5680ec7733b4672a7abadf30fe3908"} Feb 27 16:44:38 crc kubenswrapper[4814]: I0227 16:44:38.508317 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d" path="/var/lib/kubelet/pods/20695c8e-b66c-4cf1-80c4-2fc1bbd2e51d/volumes" Feb 27 16:44:38 crc kubenswrapper[4814]: I0227 16:44:38.792387 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bd1ed9cd-1472-46c3-852f-a41a2a316e40","Type":"ContainerStarted","Data":"19e9ad7dc039ddca1adf3fa62971fcde4a47044eb6b36b3668884d767e42d473"} Feb 27 16:44:39 crc kubenswrapper[4814]: I0227 16:44:39.231140 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 27 16:44:39 crc kubenswrapper[4814]: I0227 16:44:39.671464 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-845d8f46cb-qhstj" podUID="6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 27 16:44:39 crc kubenswrapper[4814]: I0227 16:44:39.804431 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bd1ed9cd-1472-46c3-852f-a41a2a316e40","Type":"ContainerStarted","Data":"2875c74e5924f4f8a7e7f106ac90c9d1b5da5a6386141049917e2b228725f36b"} Feb 27 16:44:39 crc kubenswrapper[4814]: I0227 16:44:39.810121 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ad0ff7ef-a759-4820-8c49-d971adb89ff7","Type":"ContainerStarted","Data":"f61ca20d42c6a9057741ad502eac3038b2f8eaa76bca02c692f470148d15de8a"} Feb 27 16:44:39 crc kubenswrapper[4814]: I0227 16:44:39.810486 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 16:44:39 crc kubenswrapper[4814]: I0227 16:44:39.823145 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.823130533 podStartE2EDuration="3.823130533s" podCreationTimestamp="2026-02-27 16:44:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:39.819583952 +0000 UTC m=+1292.272208782" watchObservedRunningTime="2026-02-27 16:44:39.823130533 +0000 UTC m=+1292.275755363" Feb 27 16:44:39 crc kubenswrapper[4814]: I0227 16:44:39.841993 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.964193322 podStartE2EDuration="11.841974214s" podCreationTimestamp="2026-02-27 16:44:28 +0000 UTC" firstStartedPulling="2026-02-27 16:44:29.620621308 +0000 UTC m=+1282.073246138" lastFinishedPulling="2026-02-27 16:44:38.49840221 +0000 UTC m=+1290.951027030" observedRunningTime="2026-02-27 16:44:39.836244934 +0000 UTC m=+1292.288869764" watchObservedRunningTime="2026-02-27 16:44:39.841974214 +0000 UTC m=+1292.294599034" Feb 27 16:44:41 crc kubenswrapper[4814]: I0227 16:44:41.149341 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:41 crc kubenswrapper[4814]: I0227 16:44:41.429976 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8487496c64-cv9xr" Feb 27 16:44:42 crc kubenswrapper[4814]: I0227 16:44:42.128099 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 27 16:44:42 crc kubenswrapper[4814]: I0227 16:44:42.875700 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6d4ff56d55-bj24h"] Feb 27 16:44:42 crc kubenswrapper[4814]: I0227 16:44:42.877036 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:42 crc kubenswrapper[4814]: I0227 16:44:42.900703 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 27 16:44:42 crc kubenswrapper[4814]: I0227 16:44:42.900906 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 27 16:44:42 crc kubenswrapper[4814]: I0227 16:44:42.901080 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 27 16:44:42 crc kubenswrapper[4814]: I0227 16:44:42.903601 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6d4ff56d55-bj24h"] Feb 27 16:44:42 crc kubenswrapper[4814]: I0227 16:44:42.936129 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-config-data\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:42 crc kubenswrapper[4814]: I0227 16:44:42.936179 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-run-httpd\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:42 crc kubenswrapper[4814]: I0227 16:44:42.936220 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwqqr\" (UniqueName: \"kubernetes.io/projected/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-kube-api-access-vwqqr\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:42 crc kubenswrapper[4814]: I0227 16:44:42.936247 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-public-tls-certs\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:42 crc kubenswrapper[4814]: I0227 16:44:42.936284 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-internal-tls-certs\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:42 crc kubenswrapper[4814]: I0227 16:44:42.936313 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-combined-ca-bundle\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:42 crc kubenswrapper[4814]: I0227 16:44:42.936414 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-log-httpd\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:42 crc kubenswrapper[4814]: I0227 16:44:42.936457 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-etc-swift\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.038212 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-etc-swift\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.038282 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-config-data\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.038412 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-run-httpd\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.038574 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwqqr\" (UniqueName: \"kubernetes.io/projected/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-kube-api-access-vwqqr\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.038652 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-public-tls-certs\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.038671 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-internal-tls-certs\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.038743 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-combined-ca-bundle\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.038901 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-log-httpd\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.039483 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-log-httpd\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.042896 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-run-httpd\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.047985 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-config-data\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.050419 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-internal-tls-certs\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.054828 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-etc-swift\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.060872 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-public-tls-certs\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.064990 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-combined-ca-bundle\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.066775 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwqqr\" (UniqueName: \"kubernetes.io/projected/6cc6c3ab-803a-45de-a704-2e180b3bd2ce-kube-api-access-vwqqr\") pod \"swift-proxy-6d4ff56d55-bj24h\" (UID: \"6cc6c3ab-803a-45de-a704-2e180b3bd2ce\") " pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:43 crc kubenswrapper[4814]: I0227 16:44:43.258448 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:44 crc kubenswrapper[4814]: I0227 16:44:44.803565 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:44:44 crc kubenswrapper[4814]: I0227 16:44:44.804131 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerName="ceilometer-central-agent" containerID="cri-o://99163a979e91bbcbfa552f4f52c416a375c6848873571d31e1d6b144edb2089d" gracePeriod=30 Feb 27 16:44:44 crc kubenswrapper[4814]: I0227 16:44:44.804270 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerName="proxy-httpd" containerID="cri-o://f61ca20d42c6a9057741ad502eac3038b2f8eaa76bca02c692f470148d15de8a" gracePeriod=30 Feb 27 16:44:44 crc kubenswrapper[4814]: I0227 16:44:44.804306 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerName="sg-core" containerID="cri-o://7f7b486d5cf6ef7eb75ff760e556f493eb909730b66a48113f081b82ed9e2d23" gracePeriod=30 Feb 27 16:44:44 crc kubenswrapper[4814]: I0227 16:44:44.804337 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerName="ceilometer-notification-agent" containerID="cri-o://d44aff35cb251c05ab91e5f7b2b03c03c4d744f9c8554efc9d2a35a3f56e9e5f" gracePeriod=30 Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.273862 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-lgx28"] Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.274987 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lgx28" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.293201 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lgx28"] Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.365544 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-bjcn9"] Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.367042 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bjcn9" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.374464 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-bjcn9"] Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.395131 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49d900c5-2698-4c77-8d0f-713d916ff26d-operator-scripts\") pod \"nova-api-db-create-lgx28\" (UID: \"49d900c5-2698-4c77-8d0f-713d916ff26d\") " pod="openstack/nova-api-db-create-lgx28" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.395204 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g82mz\" (UniqueName: \"kubernetes.io/projected/49d900c5-2698-4c77-8d0f-713d916ff26d-kube-api-access-g82mz\") pod \"nova-api-db-create-lgx28\" (UID: \"49d900c5-2698-4c77-8d0f-713d916ff26d\") " pod="openstack/nova-api-db-create-lgx28" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.402631 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-bb22-account-create-update-xd4zd"] Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.403826 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-bb22-account-create-update-xd4zd" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.406680 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.408808 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-bb22-account-create-update-xd4zd"] Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.470819 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-bxqnd"] Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.472010 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bxqnd" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.498836 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx7c7\" (UniqueName: \"kubernetes.io/projected/5d287b36-3c56-48b2-aaff-8f8eaadb8f50-kube-api-access-sx7c7\") pod \"nova-cell0-db-create-bjcn9\" (UID: \"5d287b36-3c56-48b2-aaff-8f8eaadb8f50\") " pod="openstack/nova-cell0-db-create-bjcn9" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.498969 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49d900c5-2698-4c77-8d0f-713d916ff26d-operator-scripts\") pod \"nova-api-db-create-lgx28\" (UID: \"49d900c5-2698-4c77-8d0f-713d916ff26d\") " pod="openstack/nova-api-db-create-lgx28" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.499048 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g82mz\" (UniqueName: \"kubernetes.io/projected/49d900c5-2698-4c77-8d0f-713d916ff26d-kube-api-access-g82mz\") pod \"nova-api-db-create-lgx28\" (UID: \"49d900c5-2698-4c77-8d0f-713d916ff26d\") " pod="openstack/nova-api-db-create-lgx28" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.499097 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d287b36-3c56-48b2-aaff-8f8eaadb8f50-operator-scripts\") pod \"nova-cell0-db-create-bjcn9\" (UID: \"5d287b36-3c56-48b2-aaff-8f8eaadb8f50\") " pod="openstack/nova-cell0-db-create-bjcn9" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.499166 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3977ccdd-8f4e-40c7-a0f8-7faa795620f0-operator-scripts\") pod \"nova-api-bb22-account-create-update-xd4zd\" (UID: \"3977ccdd-8f4e-40c7-a0f8-7faa795620f0\") " pod="openstack/nova-api-bb22-account-create-update-xd4zd" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.499205 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cgp5\" (UniqueName: \"kubernetes.io/projected/3977ccdd-8f4e-40c7-a0f8-7faa795620f0-kube-api-access-4cgp5\") pod \"nova-api-bb22-account-create-update-xd4zd\" (UID: \"3977ccdd-8f4e-40c7-a0f8-7faa795620f0\") " pod="openstack/nova-api-bb22-account-create-update-xd4zd" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.500384 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49d900c5-2698-4c77-8d0f-713d916ff26d-operator-scripts\") pod \"nova-api-db-create-lgx28\" (UID: \"49d900c5-2698-4c77-8d0f-713d916ff26d\") " pod="openstack/nova-api-db-create-lgx28" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.500725 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bxqnd"] Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.520392 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g82mz\" (UniqueName: \"kubernetes.io/projected/49d900c5-2698-4c77-8d0f-713d916ff26d-kube-api-access-g82mz\") pod \"nova-api-db-create-lgx28\" (UID: \"49d900c5-2698-4c77-8d0f-713d916ff26d\") " pod="openstack/nova-api-db-create-lgx28" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.591413 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-b173-account-create-update-n5r76"] Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.592718 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b173-account-create-update-n5r76" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.593010 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lgx28" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.593769 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-b173-account-create-update-n5r76"] Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.596461 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.600752 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfnfx\" (UniqueName: \"kubernetes.io/projected/50b9ebf8-9e8b-476f-a868-688f3439aca4-kube-api-access-gfnfx\") pod \"nova-cell1-db-create-bxqnd\" (UID: \"50b9ebf8-9e8b-476f-a868-688f3439aca4\") " pod="openstack/nova-cell1-db-create-bxqnd" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.601078 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx7c7\" (UniqueName: \"kubernetes.io/projected/5d287b36-3c56-48b2-aaff-8f8eaadb8f50-kube-api-access-sx7c7\") pod \"nova-cell0-db-create-bjcn9\" (UID: \"5d287b36-3c56-48b2-aaff-8f8eaadb8f50\") " pod="openstack/nova-cell0-db-create-bjcn9" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.601167 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50b9ebf8-9e8b-476f-a868-688f3439aca4-operator-scripts\") pod \"nova-cell1-db-create-bxqnd\" (UID: \"50b9ebf8-9e8b-476f-a868-688f3439aca4\") " pod="openstack/nova-cell1-db-create-bxqnd" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.601197 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d287b36-3c56-48b2-aaff-8f8eaadb8f50-operator-scripts\") pod \"nova-cell0-db-create-bjcn9\" (UID: \"5d287b36-3c56-48b2-aaff-8f8eaadb8f50\") " pod="openstack/nova-cell0-db-create-bjcn9" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.601259 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3977ccdd-8f4e-40c7-a0f8-7faa795620f0-operator-scripts\") pod \"nova-api-bb22-account-create-update-xd4zd\" (UID: \"3977ccdd-8f4e-40c7-a0f8-7faa795620f0\") " pod="openstack/nova-api-bb22-account-create-update-xd4zd" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.601286 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cgp5\" (UniqueName: \"kubernetes.io/projected/3977ccdd-8f4e-40c7-a0f8-7faa795620f0-kube-api-access-4cgp5\") pod \"nova-api-bb22-account-create-update-xd4zd\" (UID: \"3977ccdd-8f4e-40c7-a0f8-7faa795620f0\") " pod="openstack/nova-api-bb22-account-create-update-xd4zd" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.602143 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d287b36-3c56-48b2-aaff-8f8eaadb8f50-operator-scripts\") pod \"nova-cell0-db-create-bjcn9\" (UID: \"5d287b36-3c56-48b2-aaff-8f8eaadb8f50\") " pod="openstack/nova-cell0-db-create-bjcn9" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.602667 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3977ccdd-8f4e-40c7-a0f8-7faa795620f0-operator-scripts\") pod \"nova-api-bb22-account-create-update-xd4zd\" (UID: \"3977ccdd-8f4e-40c7-a0f8-7faa795620f0\") " pod="openstack/nova-api-bb22-account-create-update-xd4zd" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.634142 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx7c7\" (UniqueName: \"kubernetes.io/projected/5d287b36-3c56-48b2-aaff-8f8eaadb8f50-kube-api-access-sx7c7\") pod \"nova-cell0-db-create-bjcn9\" (UID: \"5d287b36-3c56-48b2-aaff-8f8eaadb8f50\") " pod="openstack/nova-cell0-db-create-bjcn9" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.638737 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cgp5\" (UniqueName: \"kubernetes.io/projected/3977ccdd-8f4e-40c7-a0f8-7faa795620f0-kube-api-access-4cgp5\") pod \"nova-api-bb22-account-create-update-xd4zd\" (UID: \"3977ccdd-8f4e-40c7-a0f8-7faa795620f0\") " pod="openstack/nova-api-bb22-account-create-update-xd4zd" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.690225 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bjcn9" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.702804 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcb4e437-b5fc-412b-9ef8-7dc917b08750-operator-scripts\") pod \"nova-cell0-b173-account-create-update-n5r76\" (UID: \"dcb4e437-b5fc-412b-9ef8-7dc917b08750\") " pod="openstack/nova-cell0-b173-account-create-update-n5r76" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.702870 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50b9ebf8-9e8b-476f-a868-688f3439aca4-operator-scripts\") pod \"nova-cell1-db-create-bxqnd\" (UID: \"50b9ebf8-9e8b-476f-a868-688f3439aca4\") " pod="openstack/nova-cell1-db-create-bxqnd" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.702904 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbkkp\" (UniqueName: \"kubernetes.io/projected/dcb4e437-b5fc-412b-9ef8-7dc917b08750-kube-api-access-fbkkp\") pod \"nova-cell0-b173-account-create-update-n5r76\" (UID: \"dcb4e437-b5fc-412b-9ef8-7dc917b08750\") " pod="openstack/nova-cell0-b173-account-create-update-n5r76" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.702980 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfnfx\" (UniqueName: \"kubernetes.io/projected/50b9ebf8-9e8b-476f-a868-688f3439aca4-kube-api-access-gfnfx\") pod \"nova-cell1-db-create-bxqnd\" (UID: \"50b9ebf8-9e8b-476f-a868-688f3439aca4\") " pod="openstack/nova-cell1-db-create-bxqnd" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.704168 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50b9ebf8-9e8b-476f-a868-688f3439aca4-operator-scripts\") pod \"nova-cell1-db-create-bxqnd\" (UID: \"50b9ebf8-9e8b-476f-a868-688f3439aca4\") " pod="openstack/nova-cell1-db-create-bxqnd" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.718054 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-bb22-account-create-update-xd4zd" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.721364 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfnfx\" (UniqueName: \"kubernetes.io/projected/50b9ebf8-9e8b-476f-a868-688f3439aca4-kube-api-access-gfnfx\") pod \"nova-cell1-db-create-bxqnd\" (UID: \"50b9ebf8-9e8b-476f-a868-688f3439aca4\") " pod="openstack/nova-cell1-db-create-bxqnd" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.776733 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-5b77-account-create-update-t2vcb"] Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.778181 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5b77-account-create-update-t2vcb" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.780069 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.802796 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5b77-account-create-update-t2vcb"] Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.804678 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcb4e437-b5fc-412b-9ef8-7dc917b08750-operator-scripts\") pod \"nova-cell0-b173-account-create-update-n5r76\" (UID: \"dcb4e437-b5fc-412b-9ef8-7dc917b08750\") " pod="openstack/nova-cell0-b173-account-create-update-n5r76" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.804774 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbkkp\" (UniqueName: \"kubernetes.io/projected/dcb4e437-b5fc-412b-9ef8-7dc917b08750-kube-api-access-fbkkp\") pod \"nova-cell0-b173-account-create-update-n5r76\" (UID: \"dcb4e437-b5fc-412b-9ef8-7dc917b08750\") " pod="openstack/nova-cell0-b173-account-create-update-n5r76" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.805352 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bxqnd" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.806646 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcb4e437-b5fc-412b-9ef8-7dc917b08750-operator-scripts\") pod \"nova-cell0-b173-account-create-update-n5r76\" (UID: \"dcb4e437-b5fc-412b-9ef8-7dc917b08750\") " pod="openstack/nova-cell0-b173-account-create-update-n5r76" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.832050 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbkkp\" (UniqueName: \"kubernetes.io/projected/dcb4e437-b5fc-412b-9ef8-7dc917b08750-kube-api-access-fbkkp\") pod \"nova-cell0-b173-account-create-update-n5r76\" (UID: \"dcb4e437-b5fc-412b-9ef8-7dc917b08750\") " pod="openstack/nova-cell0-b173-account-create-update-n5r76" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.907962 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86qwd\" (UniqueName: \"kubernetes.io/projected/80f1c20c-d371-4a0a-9c3f-e3002b086c1c-kube-api-access-86qwd\") pod \"nova-cell1-5b77-account-create-update-t2vcb\" (UID: \"80f1c20c-d371-4a0a-9c3f-e3002b086c1c\") " pod="openstack/nova-cell1-5b77-account-create-update-t2vcb" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.908092 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80f1c20c-d371-4a0a-9c3f-e3002b086c1c-operator-scripts\") pod \"nova-cell1-5b77-account-create-update-t2vcb\" (UID: \"80f1c20c-d371-4a0a-9c3f-e3002b086c1c\") " pod="openstack/nova-cell1-5b77-account-create-update-t2vcb" Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.917014 4814 generic.go:334] "Generic (PLEG): container finished" podID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerID="f61ca20d42c6a9057741ad502eac3038b2f8eaa76bca02c692f470148d15de8a" exitCode=0 Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.917058 4814 generic.go:334] "Generic (PLEG): container finished" podID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerID="7f7b486d5cf6ef7eb75ff760e556f493eb909730b66a48113f081b82ed9e2d23" exitCode=2 Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.917070 4814 generic.go:334] "Generic (PLEG): container finished" podID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerID="99163a979e91bbcbfa552f4f52c416a375c6848873571d31e1d6b144edb2089d" exitCode=0 Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.917095 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ad0ff7ef-a759-4820-8c49-d971adb89ff7","Type":"ContainerDied","Data":"f61ca20d42c6a9057741ad502eac3038b2f8eaa76bca02c692f470148d15de8a"} Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.917126 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ad0ff7ef-a759-4820-8c49-d971adb89ff7","Type":"ContainerDied","Data":"7f7b486d5cf6ef7eb75ff760e556f493eb909730b66a48113f081b82ed9e2d23"} Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.917139 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ad0ff7ef-a759-4820-8c49-d971adb89ff7","Type":"ContainerDied","Data":"99163a979e91bbcbfa552f4f52c416a375c6848873571d31e1d6b144edb2089d"} Feb 27 16:44:45 crc kubenswrapper[4814]: I0227 16:44:45.928781 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b173-account-create-update-n5r76" Feb 27 16:44:46 crc kubenswrapper[4814]: I0227 16:44:46.010193 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80f1c20c-d371-4a0a-9c3f-e3002b086c1c-operator-scripts\") pod \"nova-cell1-5b77-account-create-update-t2vcb\" (UID: \"80f1c20c-d371-4a0a-9c3f-e3002b086c1c\") " pod="openstack/nova-cell1-5b77-account-create-update-t2vcb" Feb 27 16:44:46 crc kubenswrapper[4814]: I0227 16:44:46.010369 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86qwd\" (UniqueName: \"kubernetes.io/projected/80f1c20c-d371-4a0a-9c3f-e3002b086c1c-kube-api-access-86qwd\") pod \"nova-cell1-5b77-account-create-update-t2vcb\" (UID: \"80f1c20c-d371-4a0a-9c3f-e3002b086c1c\") " pod="openstack/nova-cell1-5b77-account-create-update-t2vcb" Feb 27 16:44:46 crc kubenswrapper[4814]: I0227 16:44:46.010993 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80f1c20c-d371-4a0a-9c3f-e3002b086c1c-operator-scripts\") pod \"nova-cell1-5b77-account-create-update-t2vcb\" (UID: \"80f1c20c-d371-4a0a-9c3f-e3002b086c1c\") " pod="openstack/nova-cell1-5b77-account-create-update-t2vcb" Feb 27 16:44:46 crc kubenswrapper[4814]: I0227 16:44:46.029767 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86qwd\" (UniqueName: \"kubernetes.io/projected/80f1c20c-d371-4a0a-9c3f-e3002b086c1c-kube-api-access-86qwd\") pod \"nova-cell1-5b77-account-create-update-t2vcb\" (UID: \"80f1c20c-d371-4a0a-9c3f-e3002b086c1c\") " pod="openstack/nova-cell1-5b77-account-create-update-t2vcb" Feb 27 16:44:46 crc kubenswrapper[4814]: I0227 16:44:46.098714 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5b77-account-create-update-t2vcb" Feb 27 16:44:47 crc kubenswrapper[4814]: I0227 16:44:47.368824 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 27 16:44:47 crc kubenswrapper[4814]: I0227 16:44:47.936386 4814 generic.go:334] "Generic (PLEG): container finished" podID="131dd92c-e25a-4385-9380-84e286f09a43" containerID="d13649093c51431f472803e631c21e26b1fdd0d8c979de1fc946481976409c01" exitCode=0 Feb 27 16:44:47 crc kubenswrapper[4814]: I0227 16:44:47.936528 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85f77557cc-k22bs" event={"ID":"131dd92c-e25a-4385-9380-84e286f09a43","Type":"ContainerDied","Data":"d13649093c51431f472803e631c21e26b1fdd0d8c979de1fc946481976409c01"} Feb 27 16:44:48 crc kubenswrapper[4814]: I0227 16:44:48.947079 4814 generic.go:334] "Generic (PLEG): container finished" podID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerID="d44aff35cb251c05ab91e5f7b2b03c03c4d744f9c8554efc9d2a35a3f56e9e5f" exitCode=0 Feb 27 16:44:48 crc kubenswrapper[4814]: I0227 16:44:48.947118 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ad0ff7ef-a759-4820-8c49-d971adb89ff7","Type":"ContainerDied","Data":"d44aff35cb251c05ab91e5f7b2b03c03c4d744f9c8554efc9d2a35a3f56e9e5f"} Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.673767 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-845d8f46cb-qhstj" podUID="6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.739293 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.793545 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xs5nw\" (UniqueName: \"kubernetes.io/projected/131dd92c-e25a-4385-9380-84e286f09a43-kube-api-access-xs5nw\") pod \"131dd92c-e25a-4385-9380-84e286f09a43\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.793644 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-combined-ca-bundle\") pod \"131dd92c-e25a-4385-9380-84e286f09a43\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.793673 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-config\") pod \"131dd92c-e25a-4385-9380-84e286f09a43\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.793718 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-httpd-config\") pod \"131dd92c-e25a-4385-9380-84e286f09a43\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.793747 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-internal-tls-certs\") pod \"131dd92c-e25a-4385-9380-84e286f09a43\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.793800 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-ovndb-tls-certs\") pod \"131dd92c-e25a-4385-9380-84e286f09a43\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.793834 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-public-tls-certs\") pod \"131dd92c-e25a-4385-9380-84e286f09a43\" (UID: \"131dd92c-e25a-4385-9380-84e286f09a43\") " Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.803854 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/131dd92c-e25a-4385-9380-84e286f09a43-kube-api-access-xs5nw" (OuterVolumeSpecName: "kube-api-access-xs5nw") pod "131dd92c-e25a-4385-9380-84e286f09a43" (UID: "131dd92c-e25a-4385-9380-84e286f09a43"). InnerVolumeSpecName "kube-api-access-xs5nw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.805719 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "131dd92c-e25a-4385-9380-84e286f09a43" (UID: "131dd92c-e25a-4385-9380-84e286f09a43"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.845728 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.895166 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ad0ff7ef-a759-4820-8c49-d971adb89ff7-log-httpd\") pod \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.895468 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-combined-ca-bundle\") pod \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.895509 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-scripts\") pod \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.895588 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fcnv\" (UniqueName: \"kubernetes.io/projected/ad0ff7ef-a759-4820-8c49-d971adb89ff7-kube-api-access-4fcnv\") pod \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.895605 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-config-data\") pod \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.895703 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-sg-core-conf-yaml\") pod \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.895817 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ad0ff7ef-a759-4820-8c49-d971adb89ff7-run-httpd\") pod \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\" (UID: \"ad0ff7ef-a759-4820-8c49-d971adb89ff7\") " Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.895832 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad0ff7ef-a759-4820-8c49-d971adb89ff7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ad0ff7ef-a759-4820-8c49-d971adb89ff7" (UID: "ad0ff7ef-a759-4820-8c49-d971adb89ff7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.896681 4814 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.896703 4814 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ad0ff7ef-a759-4820-8c49-d971adb89ff7-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.896712 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xs5nw\" (UniqueName: \"kubernetes.io/projected/131dd92c-e25a-4385-9380-84e286f09a43-kube-api-access-xs5nw\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.896677 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad0ff7ef-a759-4820-8c49-d971adb89ff7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ad0ff7ef-a759-4820-8c49-d971adb89ff7" (UID: "ad0ff7ef-a759-4820-8c49-d971adb89ff7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.899485 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad0ff7ef-a759-4820-8c49-d971adb89ff7-kube-api-access-4fcnv" (OuterVolumeSpecName: "kube-api-access-4fcnv") pod "ad0ff7ef-a759-4820-8c49-d971adb89ff7" (UID: "ad0ff7ef-a759-4820-8c49-d971adb89ff7"). InnerVolumeSpecName "kube-api-access-4fcnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.900230 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-scripts" (OuterVolumeSpecName: "scripts") pod "ad0ff7ef-a759-4820-8c49-d971adb89ff7" (UID: "ad0ff7ef-a759-4820-8c49-d971adb89ff7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.908134 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "131dd92c-e25a-4385-9380-84e286f09a43" (UID: "131dd92c-e25a-4385-9380-84e286f09a43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.909578 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "131dd92c-e25a-4385-9380-84e286f09a43" (UID: "131dd92c-e25a-4385-9380-84e286f09a43"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.913338 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "131dd92c-e25a-4385-9380-84e286f09a43" (UID: "131dd92c-e25a-4385-9380-84e286f09a43"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.932815 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ad0ff7ef-a759-4820-8c49-d971adb89ff7" (UID: "ad0ff7ef-a759-4820-8c49-d971adb89ff7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.933086 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-config" (OuterVolumeSpecName: "config") pod "131dd92c-e25a-4385-9380-84e286f09a43" (UID: "131dd92c-e25a-4385-9380-84e286f09a43"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.944319 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "131dd92c-e25a-4385-9380-84e286f09a43" (UID: "131dd92c-e25a-4385-9380-84e286f09a43"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.970517 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ad0ff7ef-a759-4820-8c49-d971adb89ff7","Type":"ContainerDied","Data":"3134271f6f40fb17bfc70e878327952f1e2b51c8007c37d1b5e29524d3d37e80"} Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.970581 4814 scope.go:117] "RemoveContainer" containerID="f61ca20d42c6a9057741ad502eac3038b2f8eaa76bca02c692f470148d15de8a" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.970808 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.978770 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"cc735347-66c8-4f38-89e9-31345e59ffee","Type":"ContainerStarted","Data":"ef2cd304767df0c26ee00d6201848d04a73d4bdf5556247d157b4004bc7ec0ff"} Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.980381 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad0ff7ef-a759-4820-8c49-d971adb89ff7" (UID: "ad0ff7ef-a759-4820-8c49-d971adb89ff7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.990819 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85f77557cc-k22bs" event={"ID":"131dd92c-e25a-4385-9380-84e286f09a43","Type":"ContainerDied","Data":"f5017af6bcc2517f9c5074da11f32164cbb6bd26a0dd518f692e885f3b43d209"} Feb 27 16:44:49 crc kubenswrapper[4814]: I0227 16:44:49.990912 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85f77557cc-k22bs" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:49.998415 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:49.998454 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fcnv\" (UniqueName: \"kubernetes.io/projected/ad0ff7ef-a759-4820-8c49-d971adb89ff7-kube-api-access-4fcnv\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:49.998467 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:49.998476 4814 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:49.998487 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:49.998497 4814 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:49.998504 4814 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ad0ff7ef-a759-4820-8c49-d971adb89ff7-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:49.998514 4814 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:49.998524 4814 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/131dd92c-e25a-4385-9380-84e286f09a43-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:49.998531 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:49.998530 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.006928901 podStartE2EDuration="14.998510369s" podCreationTimestamp="2026-02-27 16:44:35 +0000 UTC" firstStartedPulling="2026-02-27 16:44:36.488873822 +0000 UTC m=+1288.941498652" lastFinishedPulling="2026-02-27 16:44:49.48045529 +0000 UTC m=+1301.933080120" observedRunningTime="2026-02-27 16:44:49.996098983 +0000 UTC m=+1302.448723813" watchObservedRunningTime="2026-02-27 16:44:49.998510369 +0000 UTC m=+1302.451135199" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.010884 4814 scope.go:117] "RemoveContainer" containerID="7f7b486d5cf6ef7eb75ff760e556f493eb909730b66a48113f081b82ed9e2d23" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.037649 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-85f77557cc-k22bs"] Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.041832 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-config-data" (OuterVolumeSpecName: "config-data") pod "ad0ff7ef-a759-4820-8c49-d971adb89ff7" (UID: "ad0ff7ef-a759-4820-8c49-d971adb89ff7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.044528 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-85f77557cc-k22bs"] Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.044954 4814 scope.go:117] "RemoveContainer" containerID="d44aff35cb251c05ab91e5f7b2b03c03c4d744f9c8554efc9d2a35a3f56e9e5f" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.100990 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad0ff7ef-a759-4820-8c49-d971adb89ff7-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.119239 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lgx28"] Feb 27 16:44:50 crc kubenswrapper[4814]: W0227 16:44:50.122651 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49d900c5_2698_4c77_8d0f_713d916ff26d.slice/crio-5c6b82d17d23d9f37d26611a5ced3fa40f40f8836134ed55a07c8bc98e255401 WatchSource:0}: Error finding container 5c6b82d17d23d9f37d26611a5ced3fa40f40f8836134ed55a07c8bc98e255401: Status 404 returned error can't find the container with id 5c6b82d17d23d9f37d26611a5ced3fa40f40f8836134ed55a07c8bc98e255401 Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.227630 4814 scope.go:117] "RemoveContainer" containerID="99163a979e91bbcbfa552f4f52c416a375c6848873571d31e1d6b144edb2089d" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.256190 4814 scope.go:117] "RemoveContainer" containerID="bdaf42af64df4a5bbcec9295585e7515e63e2138955c345374e7af4e4023a5e8" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.291934 4814 scope.go:117] "RemoveContainer" containerID="d13649093c51431f472803e631c21e26b1fdd0d8c979de1fc946481976409c01" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.315056 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.330266 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.339809 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:44:50 crc kubenswrapper[4814]: E0227 16:44:50.340173 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerName="ceilometer-notification-agent" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.340191 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerName="ceilometer-notification-agent" Feb 27 16:44:50 crc kubenswrapper[4814]: E0227 16:44:50.340202 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131dd92c-e25a-4385-9380-84e286f09a43" containerName="neutron-api" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.340209 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="131dd92c-e25a-4385-9380-84e286f09a43" containerName="neutron-api" Feb 27 16:44:50 crc kubenswrapper[4814]: E0227 16:44:50.340220 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerName="sg-core" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.340227 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerName="sg-core" Feb 27 16:44:50 crc kubenswrapper[4814]: E0227 16:44:50.340266 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerName="proxy-httpd" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.340274 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerName="proxy-httpd" Feb 27 16:44:50 crc kubenswrapper[4814]: E0227 16:44:50.340285 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131dd92c-e25a-4385-9380-84e286f09a43" containerName="neutron-httpd" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.340291 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="131dd92c-e25a-4385-9380-84e286f09a43" containerName="neutron-httpd" Feb 27 16:44:50 crc kubenswrapper[4814]: E0227 16:44:50.340303 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerName="ceilometer-central-agent" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.340309 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerName="ceilometer-central-agent" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.340475 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerName="sg-core" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.340489 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="131dd92c-e25a-4385-9380-84e286f09a43" containerName="neutron-api" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.340501 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerName="ceilometer-notification-agent" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.340514 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="131dd92c-e25a-4385-9380-84e286f09a43" containerName="neutron-httpd" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.340523 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerName="ceilometer-central-agent" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.340532 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" containerName="proxy-httpd" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.342031 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.344007 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.344271 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.361767 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.406198 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-config-data\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.406235 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b75917-fd1e-4a1d-b733-20f0d9f023dd-run-httpd\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.406283 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.406339 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.406356 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-scripts\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.406436 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b75917-fd1e-4a1d-b733-20f0d9f023dd-log-httpd\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.406468 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd2bw\" (UniqueName: \"kubernetes.io/projected/68b75917-fd1e-4a1d-b733-20f0d9f023dd-kube-api-access-bd2bw\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: W0227 16:44:50.449996 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d287b36_3c56_48b2_aaff_8f8eaadb8f50.slice/crio-afec04bb78b02c6b9e5c8f123365b4b0cdee0e9a0c713a9f5569d9e75786d16e WatchSource:0}: Error finding container afec04bb78b02c6b9e5c8f123365b4b0cdee0e9a0c713a9f5569d9e75786d16e: Status 404 returned error can't find the container with id afec04bb78b02c6b9e5c8f123365b4b0cdee0e9a0c713a9f5569d9e75786d16e Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.457389 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-bjcn9"] Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.465940 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-bb22-account-create-update-xd4zd"] Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.482878 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bxqnd"] Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.530025 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b75917-fd1e-4a1d-b733-20f0d9f023dd-log-httpd\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.530125 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd2bw\" (UniqueName: \"kubernetes.io/projected/68b75917-fd1e-4a1d-b733-20f0d9f023dd-kube-api-access-bd2bw\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.530305 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-config-data\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.530332 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b75917-fd1e-4a1d-b733-20f0d9f023dd-run-httpd\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.530389 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.530512 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.530540 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-scripts\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.530667 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b75917-fd1e-4a1d-b733-20f0d9f023dd-log-httpd\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.531597 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b75917-fd1e-4a1d-b733-20f0d9f023dd-run-httpd\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.539158 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.542089 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="131dd92c-e25a-4385-9380-84e286f09a43" path="/var/lib/kubelet/pods/131dd92c-e25a-4385-9380-84e286f09a43/volumes" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.542895 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad0ff7ef-a759-4820-8c49-d971adb89ff7" path="/var/lib/kubelet/pods/ad0ff7ef-a759-4820-8c49-d971adb89ff7/volumes" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.543764 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-config-data\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.544601 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5b77-account-create-update-t2vcb"] Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.547501 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.550822 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-scripts\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.554022 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd2bw\" (UniqueName: \"kubernetes.io/projected/68b75917-fd1e-4a1d-b733-20f0d9f023dd-kube-api-access-bd2bw\") pod \"ceilometer-0\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " pod="openstack/ceilometer-0" Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.560735 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-b173-account-create-update-n5r76"] Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.576659 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6d4ff56d55-bj24h"] Feb 27 16:44:50 crc kubenswrapper[4814]: I0227 16:44:50.686665 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.007955 4814 generic.go:334] "Generic (PLEG): container finished" podID="49d900c5-2698-4c77-8d0f-713d916ff26d" containerID="a0b6a424b296bbaed0cc99e6d686da04355e740cc540cb28cab975b8f3731f9a" exitCode=0 Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.008337 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lgx28" event={"ID":"49d900c5-2698-4c77-8d0f-713d916ff26d","Type":"ContainerDied","Data":"a0b6a424b296bbaed0cc99e6d686da04355e740cc540cb28cab975b8f3731f9a"} Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.008361 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lgx28" event={"ID":"49d900c5-2698-4c77-8d0f-713d916ff26d","Type":"ContainerStarted","Data":"5c6b82d17d23d9f37d26611a5ced3fa40f40f8836134ed55a07c8bc98e255401"} Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.011428 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b173-account-create-update-n5r76" event={"ID":"dcb4e437-b5fc-412b-9ef8-7dc917b08750","Type":"ContainerStarted","Data":"efc4ef9a6d13d0d48a72b249a5e0c945110c36e452b9bc08e44c416a32b89469"} Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.011465 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b173-account-create-update-n5r76" event={"ID":"dcb4e437-b5fc-412b-9ef8-7dc917b08750","Type":"ContainerStarted","Data":"ed8fcafeeb5206615c8c8717845b288c9c3a883fcdeef673f971638fec97c825"} Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.014193 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bxqnd" event={"ID":"50b9ebf8-9e8b-476f-a868-688f3439aca4","Type":"ContainerStarted","Data":"0b8580ebef7388d7c176924b75744c31dd6e900b6fe75afbfb66dc9414ef8eeb"} Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.014213 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bxqnd" event={"ID":"50b9ebf8-9e8b-476f-a868-688f3439aca4","Type":"ContainerStarted","Data":"d2ecc1a476c58427b244cc64a021f90b60b416f8633db730676673a4cc370f56"} Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.016948 4814 generic.go:334] "Generic (PLEG): container finished" podID="5d287b36-3c56-48b2-aaff-8f8eaadb8f50" containerID="0360f6ed9dd7bff9f9a14cba051f2c162805cc9ae3c673a435e3eb40c9f0f467" exitCode=0 Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.017003 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bjcn9" event={"ID":"5d287b36-3c56-48b2-aaff-8f8eaadb8f50","Type":"ContainerDied","Data":"0360f6ed9dd7bff9f9a14cba051f2c162805cc9ae3c673a435e3eb40c9f0f467"} Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.017023 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bjcn9" event={"ID":"5d287b36-3c56-48b2-aaff-8f8eaadb8f50","Type":"ContainerStarted","Data":"afec04bb78b02c6b9e5c8f123365b4b0cdee0e9a0c713a9f5569d9e75786d16e"} Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.020160 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6d4ff56d55-bj24h" event={"ID":"6cc6c3ab-803a-45de-a704-2e180b3bd2ce","Type":"ContainerStarted","Data":"8175cfca3b15dee17d2b7d0931ec2717686761b4ea629fc201a7473f5be49534"} Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.020195 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6d4ff56d55-bj24h" event={"ID":"6cc6c3ab-803a-45de-a704-2e180b3bd2ce","Type":"ContainerStarted","Data":"bdee49e6f309598125d912faad150554aa9f08c81ce952aa775e08684a9332b0"} Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.021062 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-bb22-account-create-update-xd4zd" event={"ID":"3977ccdd-8f4e-40c7-a0f8-7faa795620f0","Type":"ContainerStarted","Data":"7b2a20b57b5372f102f10f54662a2ca73e2f1719ebc2a26b8a8d994150912faf"} Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.021085 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-bb22-account-create-update-xd4zd" event={"ID":"3977ccdd-8f4e-40c7-a0f8-7faa795620f0","Type":"ContainerStarted","Data":"94fbd3c7ee70b1afae8cc101bb7b49b01b3e99f8e428a96df096b344da4e5439"} Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.036827 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5b77-account-create-update-t2vcb" event={"ID":"80f1c20c-d371-4a0a-9c3f-e3002b086c1c","Type":"ContainerStarted","Data":"64f19510352c40ef49f724f0e5f4340fbdf43eeba8b99ef52371658ff7008f4e"} Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.036916 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5b77-account-create-update-t2vcb" event={"ID":"80f1c20c-d371-4a0a-9c3f-e3002b086c1c","Type":"ContainerStarted","Data":"6dd44d2652f65177dbcda0d3072357a857e965a554507410b6cfdfa8ab41e244"} Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.061332 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-bxqnd" podStartSLOduration=6.06131503 podStartE2EDuration="6.06131503s" podCreationTimestamp="2026-02-27 16:44:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:51.05427324 +0000 UTC m=+1303.506898070" watchObservedRunningTime="2026-02-27 16:44:51.06131503 +0000 UTC m=+1303.513939860" Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.096271 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-b173-account-create-update-n5r76" podStartSLOduration=6.096242414 podStartE2EDuration="6.096242414s" podCreationTimestamp="2026-02-27 16:44:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:51.087851471 +0000 UTC m=+1303.540476301" watchObservedRunningTime="2026-02-27 16:44:51.096242414 +0000 UTC m=+1303.548867244" Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.106464 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-bb22-account-create-update-xd4zd" podStartSLOduration=6.106446852 podStartE2EDuration="6.106446852s" podCreationTimestamp="2026-02-27 16:44:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:51.105174443 +0000 UTC m=+1303.557799353" watchObservedRunningTime="2026-02-27 16:44:51.106446852 +0000 UTC m=+1303.559071682" Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.121385 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-5b77-account-create-update-t2vcb" podStartSLOduration=6.12137108 podStartE2EDuration="6.12137108s" podCreationTimestamp="2026-02-27 16:44:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:51.119735378 +0000 UTC m=+1303.572360208" watchObservedRunningTime="2026-02-27 16:44:51.12137108 +0000 UTC m=+1303.573995910" Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.178001 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:44:51 crc kubenswrapper[4814]: W0227 16:44:51.221282 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68b75917_fd1e_4a1d_b733_20f0d9f023dd.slice/crio-2bea676aafa2098f645468e1ed45411d5b6211e879ef0033392180cbc730cec4 WatchSource:0}: Error finding container 2bea676aafa2098f645468e1ed45411d5b6211e879ef0033392180cbc730cec4: Status 404 returned error can't find the container with id 2bea676aafa2098f645468e1ed45411d5b6211e879ef0033392180cbc730cec4 Feb 27 16:44:51 crc kubenswrapper[4814]: I0227 16:44:51.950414 4814 scope.go:117] "RemoveContainer" containerID="d0a9141a1d433581ab43efa57803ac8b74c0ad746b4ca3cd3c3b8657264681da" Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.074984 4814 generic.go:334] "Generic (PLEG): container finished" podID="3977ccdd-8f4e-40c7-a0f8-7faa795620f0" containerID="7b2a20b57b5372f102f10f54662a2ca73e2f1719ebc2a26b8a8d994150912faf" exitCode=0 Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.075076 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-bb22-account-create-update-xd4zd" event={"ID":"3977ccdd-8f4e-40c7-a0f8-7faa795620f0","Type":"ContainerDied","Data":"7b2a20b57b5372f102f10f54662a2ca73e2f1719ebc2a26b8a8d994150912faf"} Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.090288 4814 generic.go:334] "Generic (PLEG): container finished" podID="80f1c20c-d371-4a0a-9c3f-e3002b086c1c" containerID="64f19510352c40ef49f724f0e5f4340fbdf43eeba8b99ef52371658ff7008f4e" exitCode=0 Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.090621 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5b77-account-create-update-t2vcb" event={"ID":"80f1c20c-d371-4a0a-9c3f-e3002b086c1c","Type":"ContainerDied","Data":"64f19510352c40ef49f724f0e5f4340fbdf43eeba8b99ef52371658ff7008f4e"} Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.092529 4814 generic.go:334] "Generic (PLEG): container finished" podID="dcb4e437-b5fc-412b-9ef8-7dc917b08750" containerID="efc4ef9a6d13d0d48a72b249a5e0c945110c36e452b9bc08e44c416a32b89469" exitCode=0 Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.092573 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b173-account-create-update-n5r76" event={"ID":"dcb4e437-b5fc-412b-9ef8-7dc917b08750","Type":"ContainerDied","Data":"efc4ef9a6d13d0d48a72b249a5e0c945110c36e452b9bc08e44c416a32b89469"} Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.095161 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6d4ff56d55-bj24h" event={"ID":"6cc6c3ab-803a-45de-a704-2e180b3bd2ce","Type":"ContainerStarted","Data":"18d3a65d5857eff7dd543cdd3c50980e2e8c9af9464db36ae312e19098f78bef"} Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.096323 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.097319 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.109156 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b75917-fd1e-4a1d-b733-20f0d9f023dd","Type":"ContainerStarted","Data":"edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb"} Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.109542 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b75917-fd1e-4a1d-b733-20f0d9f023dd","Type":"ContainerStarted","Data":"2bea676aafa2098f645468e1ed45411d5b6211e879ef0033392180cbc730cec4"} Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.114570 4814 generic.go:334] "Generic (PLEG): container finished" podID="50b9ebf8-9e8b-476f-a868-688f3439aca4" containerID="0b8580ebef7388d7c176924b75744c31dd6e900b6fe75afbfb66dc9414ef8eeb" exitCode=0 Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.114645 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bxqnd" event={"ID":"50b9ebf8-9e8b-476f-a868-688f3439aca4","Type":"ContainerDied","Data":"0b8580ebef7388d7c176924b75744c31dd6e900b6fe75afbfb66dc9414ef8eeb"} Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.136280 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6d4ff56d55-bj24h" podStartSLOduration=10.136241341 podStartE2EDuration="10.136241341s" podCreationTimestamp="2026-02-27 16:44:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:44:52.133526097 +0000 UTC m=+1304.586150927" watchObservedRunningTime="2026-02-27 16:44:52.136241341 +0000 UTC m=+1304.588866171" Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.790947 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bjcn9" Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.796744 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lgx28" Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.894036 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sx7c7\" (UniqueName: \"kubernetes.io/projected/5d287b36-3c56-48b2-aaff-8f8eaadb8f50-kube-api-access-sx7c7\") pod \"5d287b36-3c56-48b2-aaff-8f8eaadb8f50\" (UID: \"5d287b36-3c56-48b2-aaff-8f8eaadb8f50\") " Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.894102 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d287b36-3c56-48b2-aaff-8f8eaadb8f50-operator-scripts\") pod \"5d287b36-3c56-48b2-aaff-8f8eaadb8f50\" (UID: \"5d287b36-3c56-48b2-aaff-8f8eaadb8f50\") " Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.894136 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49d900c5-2698-4c77-8d0f-713d916ff26d-operator-scripts\") pod \"49d900c5-2698-4c77-8d0f-713d916ff26d\" (UID: \"49d900c5-2698-4c77-8d0f-713d916ff26d\") " Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.894283 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g82mz\" (UniqueName: \"kubernetes.io/projected/49d900c5-2698-4c77-8d0f-713d916ff26d-kube-api-access-g82mz\") pod \"49d900c5-2698-4c77-8d0f-713d916ff26d\" (UID: \"49d900c5-2698-4c77-8d0f-713d916ff26d\") " Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.895132 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d287b36-3c56-48b2-aaff-8f8eaadb8f50-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5d287b36-3c56-48b2-aaff-8f8eaadb8f50" (UID: "5d287b36-3c56-48b2-aaff-8f8eaadb8f50"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.895873 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49d900c5-2698-4c77-8d0f-713d916ff26d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "49d900c5-2698-4c77-8d0f-713d916ff26d" (UID: "49d900c5-2698-4c77-8d0f-713d916ff26d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.899884 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49d900c5-2698-4c77-8d0f-713d916ff26d-kube-api-access-g82mz" (OuterVolumeSpecName: "kube-api-access-g82mz") pod "49d900c5-2698-4c77-8d0f-713d916ff26d" (UID: "49d900c5-2698-4c77-8d0f-713d916ff26d"). InnerVolumeSpecName "kube-api-access-g82mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.900202 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d287b36-3c56-48b2-aaff-8f8eaadb8f50-kube-api-access-sx7c7" (OuterVolumeSpecName: "kube-api-access-sx7c7") pod "5d287b36-3c56-48b2-aaff-8f8eaadb8f50" (UID: "5d287b36-3c56-48b2-aaff-8f8eaadb8f50"). InnerVolumeSpecName "kube-api-access-sx7c7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.902064 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.902119 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.996089 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g82mz\" (UniqueName: \"kubernetes.io/projected/49d900c5-2698-4c77-8d0f-713d916ff26d-kube-api-access-g82mz\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.996122 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sx7c7\" (UniqueName: \"kubernetes.io/projected/5d287b36-3c56-48b2-aaff-8f8eaadb8f50-kube-api-access-sx7c7\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.996132 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d287b36-3c56-48b2-aaff-8f8eaadb8f50-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:52 crc kubenswrapper[4814]: I0227 16:44:52.996142 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49d900c5-2698-4c77-8d0f-713d916ff26d-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.124313 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bjcn9" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.124315 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bjcn9" event={"ID":"5d287b36-3c56-48b2-aaff-8f8eaadb8f50","Type":"ContainerDied","Data":"afec04bb78b02c6b9e5c8f123365b4b0cdee0e9a0c713a9f5569d9e75786d16e"} Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.124469 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afec04bb78b02c6b9e5c8f123365b4b0cdee0e9a0c713a9f5569d9e75786d16e" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.125817 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lgx28" event={"ID":"49d900c5-2698-4c77-8d0f-713d916ff26d","Type":"ContainerDied","Data":"5c6b82d17d23d9f37d26611a5ced3fa40f40f8836134ed55a07c8bc98e255401"} Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.125845 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c6b82d17d23d9f37d26611a5ced3fa40f40f8836134ed55a07c8bc98e255401" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.125889 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lgx28" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.127500 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b75917-fd1e-4a1d-b733-20f0d9f023dd","Type":"ContainerStarted","Data":"fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2"} Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.599479 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5b77-account-create-update-t2vcb" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.687869 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-bb22-account-create-update-xd4zd" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.694286 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b173-account-create-update-n5r76" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.701656 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bxqnd" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.709660 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80f1c20c-d371-4a0a-9c3f-e3002b086c1c-operator-scripts\") pod \"80f1c20c-d371-4a0a-9c3f-e3002b086c1c\" (UID: \"80f1c20c-d371-4a0a-9c3f-e3002b086c1c\") " Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.709756 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86qwd\" (UniqueName: \"kubernetes.io/projected/80f1c20c-d371-4a0a-9c3f-e3002b086c1c-kube-api-access-86qwd\") pod \"80f1c20c-d371-4a0a-9c3f-e3002b086c1c\" (UID: \"80f1c20c-d371-4a0a-9c3f-e3002b086c1c\") " Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.711599 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80f1c20c-d371-4a0a-9c3f-e3002b086c1c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "80f1c20c-d371-4a0a-9c3f-e3002b086c1c" (UID: "80f1c20c-d371-4a0a-9c3f-e3002b086c1c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.716514 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80f1c20c-d371-4a0a-9c3f-e3002b086c1c-kube-api-access-86qwd" (OuterVolumeSpecName: "kube-api-access-86qwd") pod "80f1c20c-d371-4a0a-9c3f-e3002b086c1c" (UID: "80f1c20c-d371-4a0a-9c3f-e3002b086c1c"). InnerVolumeSpecName "kube-api-access-86qwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.811865 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbkkp\" (UniqueName: \"kubernetes.io/projected/dcb4e437-b5fc-412b-9ef8-7dc917b08750-kube-api-access-fbkkp\") pod \"dcb4e437-b5fc-412b-9ef8-7dc917b08750\" (UID: \"dcb4e437-b5fc-412b-9ef8-7dc917b08750\") " Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.812186 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfnfx\" (UniqueName: \"kubernetes.io/projected/50b9ebf8-9e8b-476f-a868-688f3439aca4-kube-api-access-gfnfx\") pod \"50b9ebf8-9e8b-476f-a868-688f3439aca4\" (UID: \"50b9ebf8-9e8b-476f-a868-688f3439aca4\") " Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.812217 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50b9ebf8-9e8b-476f-a868-688f3439aca4-operator-scripts\") pod \"50b9ebf8-9e8b-476f-a868-688f3439aca4\" (UID: \"50b9ebf8-9e8b-476f-a868-688f3439aca4\") " Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.812298 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3977ccdd-8f4e-40c7-a0f8-7faa795620f0-operator-scripts\") pod \"3977ccdd-8f4e-40c7-a0f8-7faa795620f0\" (UID: \"3977ccdd-8f4e-40c7-a0f8-7faa795620f0\") " Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.812392 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cgp5\" (UniqueName: \"kubernetes.io/projected/3977ccdd-8f4e-40c7-a0f8-7faa795620f0-kube-api-access-4cgp5\") pod \"3977ccdd-8f4e-40c7-a0f8-7faa795620f0\" (UID: \"3977ccdd-8f4e-40c7-a0f8-7faa795620f0\") " Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.812416 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcb4e437-b5fc-412b-9ef8-7dc917b08750-operator-scripts\") pod \"dcb4e437-b5fc-412b-9ef8-7dc917b08750\" (UID: \"dcb4e437-b5fc-412b-9ef8-7dc917b08750\") " Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.812744 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80f1c20c-d371-4a0a-9c3f-e3002b086c1c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.812761 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86qwd\" (UniqueName: \"kubernetes.io/projected/80f1c20c-d371-4a0a-9c3f-e3002b086c1c-kube-api-access-86qwd\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.813132 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcb4e437-b5fc-412b-9ef8-7dc917b08750-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dcb4e437-b5fc-412b-9ef8-7dc917b08750" (UID: "dcb4e437-b5fc-412b-9ef8-7dc917b08750"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.813485 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50b9ebf8-9e8b-476f-a868-688f3439aca4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "50b9ebf8-9e8b-476f-a868-688f3439aca4" (UID: "50b9ebf8-9e8b-476f-a868-688f3439aca4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.813819 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3977ccdd-8f4e-40c7-a0f8-7faa795620f0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3977ccdd-8f4e-40c7-a0f8-7faa795620f0" (UID: "3977ccdd-8f4e-40c7-a0f8-7faa795620f0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.815561 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcb4e437-b5fc-412b-9ef8-7dc917b08750-kube-api-access-fbkkp" (OuterVolumeSpecName: "kube-api-access-fbkkp") pod "dcb4e437-b5fc-412b-9ef8-7dc917b08750" (UID: "dcb4e437-b5fc-412b-9ef8-7dc917b08750"). InnerVolumeSpecName "kube-api-access-fbkkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.817274 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3977ccdd-8f4e-40c7-a0f8-7faa795620f0-kube-api-access-4cgp5" (OuterVolumeSpecName: "kube-api-access-4cgp5") pod "3977ccdd-8f4e-40c7-a0f8-7faa795620f0" (UID: "3977ccdd-8f4e-40c7-a0f8-7faa795620f0"). InnerVolumeSpecName "kube-api-access-4cgp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.825434 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50b9ebf8-9e8b-476f-a868-688f3439aca4-kube-api-access-gfnfx" (OuterVolumeSpecName: "kube-api-access-gfnfx") pod "50b9ebf8-9e8b-476f-a868-688f3439aca4" (UID: "50b9ebf8-9e8b-476f-a868-688f3439aca4"). InnerVolumeSpecName "kube-api-access-gfnfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.914780 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbkkp\" (UniqueName: \"kubernetes.io/projected/dcb4e437-b5fc-412b-9ef8-7dc917b08750-kube-api-access-fbkkp\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.914813 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfnfx\" (UniqueName: \"kubernetes.io/projected/50b9ebf8-9e8b-476f-a868-688f3439aca4-kube-api-access-gfnfx\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.914825 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50b9ebf8-9e8b-476f-a868-688f3439aca4-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.914833 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3977ccdd-8f4e-40c7-a0f8-7faa795620f0-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.914842 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cgp5\" (UniqueName: \"kubernetes.io/projected/3977ccdd-8f4e-40c7-a0f8-7faa795620f0-kube-api-access-4cgp5\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:53 crc kubenswrapper[4814]: I0227 16:44:53.914850 4814 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcb4e437-b5fc-412b-9ef8-7dc917b08750-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:54 crc kubenswrapper[4814]: I0227 16:44:54.137297 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b173-account-create-update-n5r76" event={"ID":"dcb4e437-b5fc-412b-9ef8-7dc917b08750","Type":"ContainerDied","Data":"ed8fcafeeb5206615c8c8717845b288c9c3a883fcdeef673f971638fec97c825"} Feb 27 16:44:54 crc kubenswrapper[4814]: I0227 16:44:54.137355 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed8fcafeeb5206615c8c8717845b288c9c3a883fcdeef673f971638fec97c825" Feb 27 16:44:54 crc kubenswrapper[4814]: I0227 16:44:54.137323 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b173-account-create-update-n5r76" Feb 27 16:44:54 crc kubenswrapper[4814]: I0227 16:44:54.141180 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b75917-fd1e-4a1d-b733-20f0d9f023dd","Type":"ContainerStarted","Data":"c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590"} Feb 27 16:44:54 crc kubenswrapper[4814]: I0227 16:44:54.142947 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bxqnd" event={"ID":"50b9ebf8-9e8b-476f-a868-688f3439aca4","Type":"ContainerDied","Data":"d2ecc1a476c58427b244cc64a021f90b60b416f8633db730676673a4cc370f56"} Feb 27 16:44:54 crc kubenswrapper[4814]: I0227 16:44:54.142977 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2ecc1a476c58427b244cc64a021f90b60b416f8633db730676673a4cc370f56" Feb 27 16:44:54 crc kubenswrapper[4814]: I0227 16:44:54.143028 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bxqnd" Feb 27 16:44:54 crc kubenswrapper[4814]: I0227 16:44:54.144930 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-bb22-account-create-update-xd4zd" event={"ID":"3977ccdd-8f4e-40c7-a0f8-7faa795620f0","Type":"ContainerDied","Data":"94fbd3c7ee70b1afae8cc101bb7b49b01b3e99f8e428a96df096b344da4e5439"} Feb 27 16:44:54 crc kubenswrapper[4814]: I0227 16:44:54.144964 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94fbd3c7ee70b1afae8cc101bb7b49b01b3e99f8e428a96df096b344da4e5439" Feb 27 16:44:54 crc kubenswrapper[4814]: I0227 16:44:54.144969 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-bb22-account-create-update-xd4zd" Feb 27 16:44:54 crc kubenswrapper[4814]: I0227 16:44:54.146547 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5b77-account-create-update-t2vcb" Feb 27 16:44:54 crc kubenswrapper[4814]: I0227 16:44:54.146546 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5b77-account-create-update-t2vcb" event={"ID":"80f1c20c-d371-4a0a-9c3f-e3002b086c1c","Type":"ContainerDied","Data":"6dd44d2652f65177dbcda0d3072357a857e965a554507410b6cfdfa8ab41e244"} Feb 27 16:44:54 crc kubenswrapper[4814]: I0227 16:44:54.146587 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6dd44d2652f65177dbcda0d3072357a857e965a554507410b6cfdfa8ab41e244" Feb 27 16:44:55 crc kubenswrapper[4814]: I0227 16:44:55.301682 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.251407 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jjmg8"] Feb 27 16:44:56 crc kubenswrapper[4814]: E0227 16:44:56.251754 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d900c5-2698-4c77-8d0f-713d916ff26d" containerName="mariadb-database-create" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.251765 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d900c5-2698-4c77-8d0f-713d916ff26d" containerName="mariadb-database-create" Feb 27 16:44:56 crc kubenswrapper[4814]: E0227 16:44:56.251779 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80f1c20c-d371-4a0a-9c3f-e3002b086c1c" containerName="mariadb-account-create-update" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.251785 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="80f1c20c-d371-4a0a-9c3f-e3002b086c1c" containerName="mariadb-account-create-update" Feb 27 16:44:56 crc kubenswrapper[4814]: E0227 16:44:56.251799 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b9ebf8-9e8b-476f-a868-688f3439aca4" containerName="mariadb-database-create" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.251805 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b9ebf8-9e8b-476f-a868-688f3439aca4" containerName="mariadb-database-create" Feb 27 16:44:56 crc kubenswrapper[4814]: E0227 16:44:56.251820 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3977ccdd-8f4e-40c7-a0f8-7faa795620f0" containerName="mariadb-account-create-update" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.251826 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="3977ccdd-8f4e-40c7-a0f8-7faa795620f0" containerName="mariadb-account-create-update" Feb 27 16:44:56 crc kubenswrapper[4814]: E0227 16:44:56.251838 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d287b36-3c56-48b2-aaff-8f8eaadb8f50" containerName="mariadb-database-create" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.251843 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d287b36-3c56-48b2-aaff-8f8eaadb8f50" containerName="mariadb-database-create" Feb 27 16:44:56 crc kubenswrapper[4814]: E0227 16:44:56.251853 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcb4e437-b5fc-412b-9ef8-7dc917b08750" containerName="mariadb-account-create-update" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.251859 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcb4e437-b5fc-412b-9ef8-7dc917b08750" containerName="mariadb-account-create-update" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.252023 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcb4e437-b5fc-412b-9ef8-7dc917b08750" containerName="mariadb-account-create-update" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.252035 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="3977ccdd-8f4e-40c7-a0f8-7faa795620f0" containerName="mariadb-account-create-update" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.252048 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d900c5-2698-4c77-8d0f-713d916ff26d" containerName="mariadb-database-create" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.252063 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b9ebf8-9e8b-476f-a868-688f3439aca4" containerName="mariadb-database-create" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.252070 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="80f1c20c-d371-4a0a-9c3f-e3002b086c1c" containerName="mariadb-account-create-update" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.252080 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d287b36-3c56-48b2-aaff-8f8eaadb8f50" containerName="mariadb-database-create" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.252674 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jjmg8" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.255971 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.256427 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.259246 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-x2grh" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.281078 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jjmg8"] Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.360184 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-scripts\") pod \"nova-cell0-conductor-db-sync-jjmg8\" (UID: \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\") " pod="openstack/nova-cell0-conductor-db-sync-jjmg8" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.360234 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jjmg8\" (UID: \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\") " pod="openstack/nova-cell0-conductor-db-sync-jjmg8" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.360287 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnm8l\" (UniqueName: \"kubernetes.io/projected/827ecd26-d313-4fb7-86bb-03e6e777f5a7-kube-api-access-jnm8l\") pod \"nova-cell0-conductor-db-sync-jjmg8\" (UID: \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\") " pod="openstack/nova-cell0-conductor-db-sync-jjmg8" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.360358 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-config-data\") pod \"nova-cell0-conductor-db-sync-jjmg8\" (UID: \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\") " pod="openstack/nova-cell0-conductor-db-sync-jjmg8" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.462472 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-scripts\") pod \"nova-cell0-conductor-db-sync-jjmg8\" (UID: \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\") " pod="openstack/nova-cell0-conductor-db-sync-jjmg8" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.462516 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jjmg8\" (UID: \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\") " pod="openstack/nova-cell0-conductor-db-sync-jjmg8" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.462533 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnm8l\" (UniqueName: \"kubernetes.io/projected/827ecd26-d313-4fb7-86bb-03e6e777f5a7-kube-api-access-jnm8l\") pod \"nova-cell0-conductor-db-sync-jjmg8\" (UID: \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\") " pod="openstack/nova-cell0-conductor-db-sync-jjmg8" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.462592 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-config-data\") pod \"nova-cell0-conductor-db-sync-jjmg8\" (UID: \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\") " pod="openstack/nova-cell0-conductor-db-sync-jjmg8" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.468578 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-config-data\") pod \"nova-cell0-conductor-db-sync-jjmg8\" (UID: \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\") " pod="openstack/nova-cell0-conductor-db-sync-jjmg8" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.468607 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jjmg8\" (UID: \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\") " pod="openstack/nova-cell0-conductor-db-sync-jjmg8" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.468695 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-scripts\") pod \"nova-cell0-conductor-db-sync-jjmg8\" (UID: \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\") " pod="openstack/nova-cell0-conductor-db-sync-jjmg8" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.493091 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnm8l\" (UniqueName: \"kubernetes.io/projected/827ecd26-d313-4fb7-86bb-03e6e777f5a7-kube-api-access-jnm8l\") pod \"nova-cell0-conductor-db-sync-jjmg8\" (UID: \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\") " pod="openstack/nova-cell0-conductor-db-sync-jjmg8" Feb 27 16:44:56 crc kubenswrapper[4814]: I0227 16:44:56.569589 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jjmg8" Feb 27 16:44:57 crc kubenswrapper[4814]: I0227 16:44:57.883572 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jjmg8"] Feb 27 16:44:57 crc kubenswrapper[4814]: W0227 16:44:57.890353 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod827ecd26_d313_4fb7_86bb_03e6e777f5a7.slice/crio-5eb131b727f1989859a6f39d6a1f16ad3149c04f1cbeb255d326d107afad74c6 WatchSource:0}: Error finding container 5eb131b727f1989859a6f39d6a1f16ad3149c04f1cbeb255d326d107afad74c6: Status 404 returned error can't find the container with id 5eb131b727f1989859a6f39d6a1f16ad3149c04f1cbeb255d326d107afad74c6 Feb 27 16:44:58 crc kubenswrapper[4814]: I0227 16:44:58.195402 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b75917-fd1e-4a1d-b733-20f0d9f023dd","Type":"ContainerStarted","Data":"30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e"} Feb 27 16:44:58 crc kubenswrapper[4814]: I0227 16:44:58.196840 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 16:44:58 crc kubenswrapper[4814]: I0227 16:44:58.196940 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerName="proxy-httpd" containerID="cri-o://30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e" gracePeriod=30 Feb 27 16:44:58 crc kubenswrapper[4814]: I0227 16:44:58.197068 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerName="sg-core" containerID="cri-o://c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590" gracePeriod=30 Feb 27 16:44:58 crc kubenswrapper[4814]: I0227 16:44:58.197157 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerName="ceilometer-notification-agent" containerID="cri-o://fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2" gracePeriod=30 Feb 27 16:44:58 crc kubenswrapper[4814]: I0227 16:44:58.197406 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerName="ceilometer-central-agent" containerID="cri-o://edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb" gracePeriod=30 Feb 27 16:44:58 crc kubenswrapper[4814]: I0227 16:44:58.197661 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jjmg8" event={"ID":"827ecd26-d313-4fb7-86bb-03e6e777f5a7","Type":"ContainerStarted","Data":"5eb131b727f1989859a6f39d6a1f16ad3149c04f1cbeb255d326d107afad74c6"} Feb 27 16:44:58 crc kubenswrapper[4814]: I0227 16:44:58.252956 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.103289509 podStartE2EDuration="8.252934334s" podCreationTimestamp="2026-02-27 16:44:50 +0000 UTC" firstStartedPulling="2026-02-27 16:44:51.225868058 +0000 UTC m=+1303.678492888" lastFinishedPulling="2026-02-27 16:44:57.375512883 +0000 UTC m=+1309.828137713" observedRunningTime="2026-02-27 16:44:58.250498268 +0000 UTC m=+1310.703123098" watchObservedRunningTime="2026-02-27 16:44:58.252934334 +0000 UTC m=+1310.705559174" Feb 27 16:44:58 crc kubenswrapper[4814]: I0227 16:44:58.268619 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:58 crc kubenswrapper[4814]: I0227 16:44:58.272423 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6d4ff56d55-bj24h" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.179622 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.211763 4814 generic.go:334] "Generic (PLEG): container finished" podID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerID="30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e" exitCode=0 Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.211794 4814 generic.go:334] "Generic (PLEG): container finished" podID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerID="c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590" exitCode=2 Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.211801 4814 generic.go:334] "Generic (PLEG): container finished" podID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerID="fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2" exitCode=0 Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.211807 4814 generic.go:334] "Generic (PLEG): container finished" podID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerID="edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb" exitCode=0 Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.211904 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.212177 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b75917-fd1e-4a1d-b733-20f0d9f023dd","Type":"ContainerDied","Data":"30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e"} Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.212221 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b75917-fd1e-4a1d-b733-20f0d9f023dd","Type":"ContainerDied","Data":"c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590"} Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.212232 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b75917-fd1e-4a1d-b733-20f0d9f023dd","Type":"ContainerDied","Data":"fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2"} Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.212241 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b75917-fd1e-4a1d-b733-20f0d9f023dd","Type":"ContainerDied","Data":"edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb"} Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.212279 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b75917-fd1e-4a1d-b733-20f0d9f023dd","Type":"ContainerDied","Data":"2bea676aafa2098f645468e1ed45411d5b6211e879ef0033392180cbc730cec4"} Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.212295 4814 scope.go:117] "RemoveContainer" containerID="30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.248525 4814 scope.go:117] "RemoveContainer" containerID="c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.292230 4814 scope.go:117] "RemoveContainer" containerID="fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.323185 4814 scope.go:117] "RemoveContainer" containerID="edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.352883 4814 scope.go:117] "RemoveContainer" containerID="30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e" Feb 27 16:44:59 crc kubenswrapper[4814]: E0227 16:44:59.353289 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e\": container with ID starting with 30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e not found: ID does not exist" containerID="30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.353324 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e"} err="failed to get container status \"30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e\": rpc error: code = NotFound desc = could not find container \"30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e\": container with ID starting with 30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e not found: ID does not exist" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.353348 4814 scope.go:117] "RemoveContainer" containerID="c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590" Feb 27 16:44:59 crc kubenswrapper[4814]: E0227 16:44:59.353647 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590\": container with ID starting with c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590 not found: ID does not exist" containerID="c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.353666 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590"} err="failed to get container status \"c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590\": rpc error: code = NotFound desc = could not find container \"c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590\": container with ID starting with c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590 not found: ID does not exist" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.353681 4814 scope.go:117] "RemoveContainer" containerID="fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2" Feb 27 16:44:59 crc kubenswrapper[4814]: E0227 16:44:59.353879 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2\": container with ID starting with fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2 not found: ID does not exist" containerID="fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.353898 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2"} err="failed to get container status \"fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2\": rpc error: code = NotFound desc = could not find container \"fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2\": container with ID starting with fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2 not found: ID does not exist" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.353917 4814 scope.go:117] "RemoveContainer" containerID="edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb" Feb 27 16:44:59 crc kubenswrapper[4814]: E0227 16:44:59.354313 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb\": container with ID starting with edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb not found: ID does not exist" containerID="edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.354336 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb"} err="failed to get container status \"edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb\": rpc error: code = NotFound desc = could not find container \"edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb\": container with ID starting with edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb not found: ID does not exist" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.354348 4814 scope.go:117] "RemoveContainer" containerID="30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.354523 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e"} err="failed to get container status \"30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e\": rpc error: code = NotFound desc = could not find container \"30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e\": container with ID starting with 30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e not found: ID does not exist" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.354538 4814 scope.go:117] "RemoveContainer" containerID="c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.354713 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590"} err="failed to get container status \"c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590\": rpc error: code = NotFound desc = could not find container \"c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590\": container with ID starting with c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590 not found: ID does not exist" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.354730 4814 scope.go:117] "RemoveContainer" containerID="fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.354973 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2"} err="failed to get container status \"fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2\": rpc error: code = NotFound desc = could not find container \"fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2\": container with ID starting with fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2 not found: ID does not exist" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.354992 4814 scope.go:117] "RemoveContainer" containerID="edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.355246 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb"} err="failed to get container status \"edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb\": rpc error: code = NotFound desc = could not find container \"edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb\": container with ID starting with edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb not found: ID does not exist" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.355348 4814 scope.go:117] "RemoveContainer" containerID="30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.355690 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e"} err="failed to get container status \"30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e\": rpc error: code = NotFound desc = could not find container \"30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e\": container with ID starting with 30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e not found: ID does not exist" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.355779 4814 scope.go:117] "RemoveContainer" containerID="c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.355998 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-sg-core-conf-yaml\") pod \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.356041 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-combined-ca-bundle\") pod \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.356098 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b75917-fd1e-4a1d-b733-20f0d9f023dd-log-httpd\") pod \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.356174 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-config-data\") pod \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.356266 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b75917-fd1e-4a1d-b733-20f0d9f023dd-run-httpd\") pod \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.356336 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-scripts\") pod \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.356370 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bd2bw\" (UniqueName: \"kubernetes.io/projected/68b75917-fd1e-4a1d-b733-20f0d9f023dd-kube-api-access-bd2bw\") pod \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.357503 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590"} err="failed to get container status \"c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590\": rpc error: code = NotFound desc = could not find container \"c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590\": container with ID starting with c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590 not found: ID does not exist" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.357561 4814 scope.go:117] "RemoveContainer" containerID="fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.358278 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68b75917-fd1e-4a1d-b733-20f0d9f023dd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "68b75917-fd1e-4a1d-b733-20f0d9f023dd" (UID: "68b75917-fd1e-4a1d-b733-20f0d9f023dd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.358497 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68b75917-fd1e-4a1d-b733-20f0d9f023dd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "68b75917-fd1e-4a1d-b733-20f0d9f023dd" (UID: "68b75917-fd1e-4a1d-b733-20f0d9f023dd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.358041 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2"} err="failed to get container status \"fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2\": rpc error: code = NotFound desc = could not find container \"fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2\": container with ID starting with fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2 not found: ID does not exist" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.358541 4814 scope.go:117] "RemoveContainer" containerID="edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.359577 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb"} err="failed to get container status \"edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb\": rpc error: code = NotFound desc = could not find container \"edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb\": container with ID starting with edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb not found: ID does not exist" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.359785 4814 scope.go:117] "RemoveContainer" containerID="30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.361236 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e"} err="failed to get container status \"30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e\": rpc error: code = NotFound desc = could not find container \"30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e\": container with ID starting with 30c9939fbdec114f05d358250a36e95599d1ee974c6360eb0c41fc8bfb51db0e not found: ID does not exist" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.361297 4814 scope.go:117] "RemoveContainer" containerID="c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.362059 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590"} err="failed to get container status \"c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590\": rpc error: code = NotFound desc = could not find container \"c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590\": container with ID starting with c8f8adf13d92bdb22dee4549c11885cfe77d529f9a20b27fee83681ed4914590 not found: ID does not exist" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.362125 4814 scope.go:117] "RemoveContainer" containerID="fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.362429 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2"} err="failed to get container status \"fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2\": rpc error: code = NotFound desc = could not find container \"fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2\": container with ID starting with fab830e8f72b69db5a580ad8c9e7a6fd46d84b8424c0d2444efb7bd99649cff2 not found: ID does not exist" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.362456 4814 scope.go:117] "RemoveContainer" containerID="edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.362616 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb"} err="failed to get container status \"edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb\": rpc error: code = NotFound desc = could not find container \"edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb\": container with ID starting with edeaf36ad10d6c57892722c3bf10ad2a356e4096895ac301b09cc2f5dd33d6eb not found: ID does not exist" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.366166 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68b75917-fd1e-4a1d-b733-20f0d9f023dd-kube-api-access-bd2bw" (OuterVolumeSpecName: "kube-api-access-bd2bw") pod "68b75917-fd1e-4a1d-b733-20f0d9f023dd" (UID: "68b75917-fd1e-4a1d-b733-20f0d9f023dd"). InnerVolumeSpecName "kube-api-access-bd2bw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.368544 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-scripts" (OuterVolumeSpecName: "scripts") pod "68b75917-fd1e-4a1d-b733-20f0d9f023dd" (UID: "68b75917-fd1e-4a1d-b733-20f0d9f023dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.398067 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "68b75917-fd1e-4a1d-b733-20f0d9f023dd" (UID: "68b75917-fd1e-4a1d-b733-20f0d9f023dd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.457856 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68b75917-fd1e-4a1d-b733-20f0d9f023dd" (UID: "68b75917-fd1e-4a1d-b733-20f0d9f023dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.458276 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-combined-ca-bundle\") pod \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\" (UID: \"68b75917-fd1e-4a1d-b733-20f0d9f023dd\") " Feb 27 16:44:59 crc kubenswrapper[4814]: W0227 16:44:59.458425 4814 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/68b75917-fd1e-4a1d-b733-20f0d9f023dd/volumes/kubernetes.io~secret/combined-ca-bundle Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.458448 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68b75917-fd1e-4a1d-b733-20f0d9f023dd" (UID: "68b75917-fd1e-4a1d-b733-20f0d9f023dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.458673 4814 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b75917-fd1e-4a1d-b733-20f0d9f023dd-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.458686 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.458695 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bd2bw\" (UniqueName: \"kubernetes.io/projected/68b75917-fd1e-4a1d-b733-20f0d9f023dd-kube-api-access-bd2bw\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.458704 4814 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.458712 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.458720 4814 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b75917-fd1e-4a1d-b733-20f0d9f023dd-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.485830 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-config-data" (OuterVolumeSpecName: "config-data") pod "68b75917-fd1e-4a1d-b733-20f0d9f023dd" (UID: "68b75917-fd1e-4a1d-b733-20f0d9f023dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.560000 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68b75917-fd1e-4a1d-b733-20f0d9f023dd-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.566400 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.581193 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.600925 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:44:59 crc kubenswrapper[4814]: E0227 16:44:59.601394 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerName="sg-core" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.601411 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerName="sg-core" Feb 27 16:44:59 crc kubenswrapper[4814]: E0227 16:44:59.601436 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerName="ceilometer-central-agent" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.601448 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerName="ceilometer-central-agent" Feb 27 16:44:59 crc kubenswrapper[4814]: E0227 16:44:59.601466 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerName="ceilometer-notification-agent" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.601475 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerName="ceilometer-notification-agent" Feb 27 16:44:59 crc kubenswrapper[4814]: E0227 16:44:59.601491 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerName="proxy-httpd" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.601505 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerName="proxy-httpd" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.601676 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerName="proxy-httpd" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.601695 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerName="ceilometer-central-agent" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.601708 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerName="sg-core" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.601724 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" containerName="ceilometer-notification-agent" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.603425 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.605893 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.606061 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.609609 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.671187 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-845d8f46cb-qhstj" podUID="6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.671624 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.763689 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b665060-ed5a-4de3-91bb-606549a98922-run-httpd\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.763774 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.763805 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdfrv\" (UniqueName: \"kubernetes.io/projected/5b665060-ed5a-4de3-91bb-606549a98922-kube-api-access-pdfrv\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.763872 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.763909 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-config-data\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.763927 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-scripts\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.763954 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b665060-ed5a-4de3-91bb-606549a98922-log-httpd\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.865509 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b665060-ed5a-4de3-91bb-606549a98922-run-httpd\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.865585 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.865609 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdfrv\" (UniqueName: \"kubernetes.io/projected/5b665060-ed5a-4de3-91bb-606549a98922-kube-api-access-pdfrv\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.865659 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.865690 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-config-data\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.865706 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-scripts\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.865725 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b665060-ed5a-4de3-91bb-606549a98922-log-httpd\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.865943 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b665060-ed5a-4de3-91bb-606549a98922-run-httpd\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.866189 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b665060-ed5a-4de3-91bb-606549a98922-log-httpd\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.870468 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.870816 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.875349 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-config-data\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.875766 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-scripts\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.882560 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdfrv\" (UniqueName: \"kubernetes.io/projected/5b665060-ed5a-4de3-91bb-606549a98922-kube-api-access-pdfrv\") pod \"ceilometer-0\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " pod="openstack/ceilometer-0" Feb 27 16:44:59 crc kubenswrapper[4814]: I0227 16:44:59.918576 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.137292 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq"] Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.139874 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.142556 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.142822 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.157488 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq"] Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.274733 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa51609e-188a-4ded-962b-9f9641404afc-secret-volume\") pod \"collect-profiles-29536845-zzkrq\" (UID: \"aa51609e-188a-4ded-962b-9f9641404afc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.274811 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl9hk\" (UniqueName: \"kubernetes.io/projected/aa51609e-188a-4ded-962b-9f9641404afc-kube-api-access-sl9hk\") pod \"collect-profiles-29536845-zzkrq\" (UID: \"aa51609e-188a-4ded-962b-9f9641404afc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.274892 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa51609e-188a-4ded-962b-9f9641404afc-config-volume\") pod \"collect-profiles-29536845-zzkrq\" (UID: \"aa51609e-188a-4ded-962b-9f9641404afc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.367929 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:45:00 crc kubenswrapper[4814]: W0227 16:45:00.375980 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b665060_ed5a_4de3_91bb_606549a98922.slice/crio-14031101bff67b61320750949fc6a6456c79ca35cf5f608171bd0d63cc9c6b93 WatchSource:0}: Error finding container 14031101bff67b61320750949fc6a6456c79ca35cf5f608171bd0d63cc9c6b93: Status 404 returned error can't find the container with id 14031101bff67b61320750949fc6a6456c79ca35cf5f608171bd0d63cc9c6b93 Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.376067 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa51609e-188a-4ded-962b-9f9641404afc-config-volume\") pod \"collect-profiles-29536845-zzkrq\" (UID: \"aa51609e-188a-4ded-962b-9f9641404afc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.376203 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa51609e-188a-4ded-962b-9f9641404afc-secret-volume\") pod \"collect-profiles-29536845-zzkrq\" (UID: \"aa51609e-188a-4ded-962b-9f9641404afc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.376285 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl9hk\" (UniqueName: \"kubernetes.io/projected/aa51609e-188a-4ded-962b-9f9641404afc-kube-api-access-sl9hk\") pod \"collect-profiles-29536845-zzkrq\" (UID: \"aa51609e-188a-4ded-962b-9f9641404afc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.377334 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa51609e-188a-4ded-962b-9f9641404afc-config-volume\") pod \"collect-profiles-29536845-zzkrq\" (UID: \"aa51609e-188a-4ded-962b-9f9641404afc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.382737 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa51609e-188a-4ded-962b-9f9641404afc-secret-volume\") pod \"collect-profiles-29536845-zzkrq\" (UID: \"aa51609e-188a-4ded-962b-9f9641404afc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.394068 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl9hk\" (UniqueName: \"kubernetes.io/projected/aa51609e-188a-4ded-962b-9f9641404afc-kube-api-access-sl9hk\") pod \"collect-profiles-29536845-zzkrq\" (UID: \"aa51609e-188a-4ded-962b-9f9641404afc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.459157 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.500835 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68b75917-fd1e-4a1d-b733-20f0d9f023dd" path="/var/lib/kubelet/pods/68b75917-fd1e-4a1d-b733-20f0d9f023dd/volumes" Feb 27 16:45:00 crc kubenswrapper[4814]: I0227 16:45:00.893523 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq"] Feb 27 16:45:00 crc kubenswrapper[4814]: W0227 16:45:00.904167 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa51609e_188a_4ded_962b_9f9641404afc.slice/crio-eb7a32b82b3fba9fb45c72f26ffdb9697916b2788f479c2eb70fe921070b3abe WatchSource:0}: Error finding container eb7a32b82b3fba9fb45c72f26ffdb9697916b2788f479c2eb70fe921070b3abe: Status 404 returned error can't find the container with id eb7a32b82b3fba9fb45c72f26ffdb9697916b2788f479c2eb70fe921070b3abe Feb 27 16:45:01 crc kubenswrapper[4814]: I0227 16:45:01.237447 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" event={"ID":"aa51609e-188a-4ded-962b-9f9641404afc","Type":"ContainerStarted","Data":"a7fbeff9a2f69b9ac82fe3541fa4c702b7fc9c91b28f6827d4bef735d3c13471"} Feb 27 16:45:01 crc kubenswrapper[4814]: I0227 16:45:01.237717 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" event={"ID":"aa51609e-188a-4ded-962b-9f9641404afc","Type":"ContainerStarted","Data":"eb7a32b82b3fba9fb45c72f26ffdb9697916b2788f479c2eb70fe921070b3abe"} Feb 27 16:45:01 crc kubenswrapper[4814]: I0227 16:45:01.240108 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b665060-ed5a-4de3-91bb-606549a98922","Type":"ContainerStarted","Data":"5b43c486e1f1b2d7fc67d3724f185be74357fc71f3f713b6333cf59df95e5a60"} Feb 27 16:45:01 crc kubenswrapper[4814]: I0227 16:45:01.240137 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b665060-ed5a-4de3-91bb-606549a98922","Type":"ContainerStarted","Data":"14031101bff67b61320750949fc6a6456c79ca35cf5f608171bd0d63cc9c6b93"} Feb 27 16:45:01 crc kubenswrapper[4814]: I0227 16:45:01.261426 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" podStartSLOduration=1.26139688 podStartE2EDuration="1.26139688s" podCreationTimestamp="2026-02-27 16:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:45:01.255890767 +0000 UTC m=+1313.708515597" watchObservedRunningTime="2026-02-27 16:45:01.26139688 +0000 UTC m=+1313.714021710" Feb 27 16:45:01 crc kubenswrapper[4814]: I0227 16:45:01.455911 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 16:45:01 crc kubenswrapper[4814]: I0227 16:45:01.456979 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" containerName="glance-log" containerID="cri-o://0c9eb449398eb3654fc7114dbf4cf0886a0c7ec6574f29d26e066a62c7c64417" gracePeriod=30 Feb 27 16:45:01 crc kubenswrapper[4814]: I0227 16:45:01.457135 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" containerName="glance-httpd" containerID="cri-o://d407574566976064af8e3d96a78617ed018b92a07e84a6d1ddc32338af3c2c3e" gracePeriod=30 Feb 27 16:45:02 crc kubenswrapper[4814]: I0227 16:45:02.251946 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b665060-ed5a-4de3-91bb-606549a98922","Type":"ContainerStarted","Data":"0e2ee7f214497e2ad846633511393a03afe7c675c6b16de11be23b43922a86e9"} Feb 27 16:45:02 crc kubenswrapper[4814]: I0227 16:45:02.255344 4814 generic.go:334] "Generic (PLEG): container finished" podID="2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" containerID="0c9eb449398eb3654fc7114dbf4cf0886a0c7ec6574f29d26e066a62c7c64417" exitCode=143 Feb 27 16:45:02 crc kubenswrapper[4814]: I0227 16:45:02.255397 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5","Type":"ContainerDied","Data":"0c9eb449398eb3654fc7114dbf4cf0886a0c7ec6574f29d26e066a62c7c64417"} Feb 27 16:45:02 crc kubenswrapper[4814]: I0227 16:45:02.258075 4814 generic.go:334] "Generic (PLEG): container finished" podID="aa51609e-188a-4ded-962b-9f9641404afc" containerID="a7fbeff9a2f69b9ac82fe3541fa4c702b7fc9c91b28f6827d4bef735d3c13471" exitCode=0 Feb 27 16:45:02 crc kubenswrapper[4814]: I0227 16:45:02.258108 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" event={"ID":"aa51609e-188a-4ded-962b-9f9641404afc","Type":"ContainerDied","Data":"a7fbeff9a2f69b9ac82fe3541fa4c702b7fc9c91b28f6827d4bef735d3c13471"} Feb 27 16:45:02 crc kubenswrapper[4814]: I0227 16:45:02.548628 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:45:04 crc kubenswrapper[4814]: I0227 16:45:04.279965 4814 generic.go:334] "Generic (PLEG): container finished" podID="6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" containerID="19807c67c77d45f3562d439c93d9fa848a39b696dec9481c4ec2a11a35968141" exitCode=137 Feb 27 16:45:04 crc kubenswrapper[4814]: I0227 16:45:04.280053 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-845d8f46cb-qhstj" event={"ID":"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1","Type":"ContainerDied","Data":"19807c67c77d45f3562d439c93d9fa848a39b696dec9481c4ec2a11a35968141"} Feb 27 16:45:05 crc kubenswrapper[4814]: I0227 16:45:05.291768 4814 generic.go:334] "Generic (PLEG): container finished" podID="2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" containerID="d407574566976064af8e3d96a78617ed018b92a07e84a6d1ddc32338af3c2c3e" exitCode=0 Feb 27 16:45:05 crc kubenswrapper[4814]: I0227 16:45:05.291858 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5","Type":"ContainerDied","Data":"d407574566976064af8e3d96a78617ed018b92a07e84a6d1ddc32338af3c2c3e"} Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.310910 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.342601 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.367587 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" event={"ID":"aa51609e-188a-4ded-962b-9f9641404afc","Type":"ContainerDied","Data":"eb7a32b82b3fba9fb45c72f26ffdb9697916b2788f479c2eb70fe921070b3abe"} Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.367641 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb7a32b82b3fba9fb45c72f26ffdb9697916b2788f479c2eb70fe921070b3abe" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.367723 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.435233 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa51609e-188a-4ded-962b-9f9641404afc-secret-volume\") pod \"aa51609e-188a-4ded-962b-9f9641404afc\" (UID: \"aa51609e-188a-4ded-962b-9f9641404afc\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.435295 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa51609e-188a-4ded-962b-9f9641404afc-config-volume\") pod \"aa51609e-188a-4ded-962b-9f9641404afc\" (UID: \"aa51609e-188a-4ded-962b-9f9641404afc\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.435328 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-logs\") pod \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.435366 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-config-data\") pod \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.435391 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-scripts\") pod \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.435508 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-combined-ca-bundle\") pod \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.435567 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-horizon-secret-key\") pod \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.435620 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-horizon-tls-certs\") pod \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.435655 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv9vm\" (UniqueName: \"kubernetes.io/projected/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-kube-api-access-bv9vm\") pod \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\" (UID: \"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.435697 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sl9hk\" (UniqueName: \"kubernetes.io/projected/aa51609e-188a-4ded-962b-9f9641404afc-kube-api-access-sl9hk\") pod \"aa51609e-188a-4ded-962b-9f9641404afc\" (UID: \"aa51609e-188a-4ded-962b-9f9641404afc\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.452932 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa51609e-188a-4ded-962b-9f9641404afc-kube-api-access-sl9hk" (OuterVolumeSpecName: "kube-api-access-sl9hk") pod "aa51609e-188a-4ded-962b-9f9641404afc" (UID: "aa51609e-188a-4ded-962b-9f9641404afc"). InnerVolumeSpecName "kube-api-access-sl9hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.457363 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa51609e-188a-4ded-962b-9f9641404afc-config-volume" (OuterVolumeSpecName: "config-volume") pod "aa51609e-188a-4ded-962b-9f9641404afc" (UID: "aa51609e-188a-4ded-962b-9f9641404afc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.472774 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-845d8f46cb-qhstj" event={"ID":"6c9011a2-17bc-42ba-a6cf-13ec98cd75b1","Type":"ContainerDied","Data":"8490be3507800d70feb1b18bf4bbe755740d3013e1f131ad3df196e1c73cec7b"} Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.472825 4814 scope.go:117] "RemoveContainer" containerID="5c624a9feb184d9603fa9c3597fae610b34d3632ba8c53b56374223a4f027fd2" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.472982 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-845d8f46cb-qhstj" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.495919 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-logs" (OuterVolumeSpecName: "logs") pod "6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" (UID: "6c9011a2-17bc-42ba-a6cf-13ec98cd75b1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.520222 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" (UID: "6c9011a2-17bc-42ba-a6cf-13ec98cd75b1"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.521062 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-kube-api-access-bv9vm" (OuterVolumeSpecName: "kube-api-access-bv9vm") pod "6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" (UID: "6c9011a2-17bc-42ba-a6cf-13ec98cd75b1"). InnerVolumeSpecName "kube-api-access-bv9vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.525232 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa51609e-188a-4ded-962b-9f9641404afc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "aa51609e-188a-4ded-962b-9f9641404afc" (UID: "aa51609e-188a-4ded-962b-9f9641404afc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.538481 4814 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.538507 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv9vm\" (UniqueName: \"kubernetes.io/projected/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-kube-api-access-bv9vm\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.538519 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sl9hk\" (UniqueName: \"kubernetes.io/projected/aa51609e-188a-4ded-962b-9f9641404afc-kube-api-access-sl9hk\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.538527 4814 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa51609e-188a-4ded-962b-9f9641404afc-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.538536 4814 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa51609e-188a-4ded-962b-9f9641404afc-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.538544 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.598784 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-scripts" (OuterVolumeSpecName: "scripts") pod "6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" (UID: "6c9011a2-17bc-42ba-a6cf-13ec98cd75b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.640535 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.656513 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-config-data" (OuterVolumeSpecName: "config-data") pod "6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" (UID: "6c9011a2-17bc-42ba-a6cf-13ec98cd75b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.675901 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" (UID: "6c9011a2-17bc-42ba-a6cf-13ec98cd75b1"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.678380 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" (UID: "6c9011a2-17bc-42ba-a6cf-13ec98cd75b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.742211 4814 scope.go:117] "RemoveContainer" containerID="19807c67c77d45f3562d439c93d9fa848a39b696dec9481c4ec2a11a35968141" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.742669 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.742705 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.742718 4814 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.839171 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.842321 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-845d8f46cb-qhstj"] Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.849167 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-845d8f46cb-qhstj"] Feb 27 16:45:08 crc kubenswrapper[4814]: E0227 16:45:08.913829 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa51609e_188a_4ded_962b_9f9641404afc.slice/crio-eb7a32b82b3fba9fb45c72f26ffdb9697916b2788f479c2eb70fe921070b3abe\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa51609e_188a_4ded_962b_9f9641404afc.slice\": RecentStats: unable to find data in memory cache]" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.948045 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.948187 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-combined-ca-bundle\") pod \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.948231 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdqtd\" (UniqueName: \"kubernetes.io/projected/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-kube-api-access-zdqtd\") pod \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.948258 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-config-data\") pod \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.948344 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-httpd-run\") pod \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.948370 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-scripts\") pod \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.948397 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-internal-tls-certs\") pod \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.948423 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-logs\") pod \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\" (UID: \"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5\") " Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.949302 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-logs" (OuterVolumeSpecName: "logs") pod "2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" (UID: "2f8e8e66-aaf0-43d6-b990-e046d17f2fb5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.950779 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" (UID: "2f8e8e66-aaf0-43d6-b990-e046d17f2fb5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.952868 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" (UID: "2f8e8e66-aaf0-43d6-b990-e046d17f2fb5"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.953461 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-scripts" (OuterVolumeSpecName: "scripts") pod "2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" (UID: "2f8e8e66-aaf0-43d6-b990-e046d17f2fb5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.963404 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-kube-api-access-zdqtd" (OuterVolumeSpecName: "kube-api-access-zdqtd") pod "2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" (UID: "2f8e8e66-aaf0-43d6-b990-e046d17f2fb5"). InnerVolumeSpecName "kube-api-access-zdqtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:45:08 crc kubenswrapper[4814]: I0227 16:45:08.997508 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" (UID: "2f8e8e66-aaf0-43d6-b990-e046d17f2fb5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.013265 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-config-data" (OuterVolumeSpecName: "config-data") pod "2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" (UID: "2f8e8e66-aaf0-43d6-b990-e046d17f2fb5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.048033 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" (UID: "2f8e8e66-aaf0-43d6-b990-e046d17f2fb5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.049988 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.050012 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdqtd\" (UniqueName: \"kubernetes.io/projected/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-kube-api-access-zdqtd\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.050027 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.050035 4814 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.050045 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.050053 4814 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.050060 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.050082 4814 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.069178 4814 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.151285 4814 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.355166 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.355421 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" containerName="glance-log" containerID="cri-o://6f9d88326b4353e59265198023144b951d9a4aad7e4934d48508f34f04a52792" gracePeriod=30 Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.355579 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" containerName="glance-httpd" containerID="cri-o://04714ba3748af9f5fe9f795ce59413df20029c8ea6f5ebf98fe3748e51f45baf" gracePeriod=30 Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.536999 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jjmg8" event={"ID":"827ecd26-d313-4fb7-86bb-03e6e777f5a7","Type":"ContainerStarted","Data":"5f2cc4f12adcccdec8db26f18084d239bc8397263cda448b3eb475afe5553635"} Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.545844 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b665060-ed5a-4de3-91bb-606549a98922","Type":"ContainerStarted","Data":"fb989c51c54614706dad712ee8a0ca4d242666e2c89b1ef2f95e44d335b144ab"} Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.547915 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2f8e8e66-aaf0-43d6-b990-e046d17f2fb5","Type":"ContainerDied","Data":"82c9148f20c6fdb72bfaae8d94beaf97dd4c55f9689a736c2fcd251db6504b97"} Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.548176 4814 scope.go:117] "RemoveContainer" containerID="d407574566976064af8e3d96a78617ed018b92a07e84a6d1ddc32338af3c2c3e" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.548689 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.554579 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-jjmg8" podStartSLOduration=3.259183166 podStartE2EDuration="13.554551397s" podCreationTimestamp="2026-02-27 16:44:56 +0000 UTC" firstStartedPulling="2026-02-27 16:44:57.892527118 +0000 UTC m=+1310.345151948" lastFinishedPulling="2026-02-27 16:45:08.187895349 +0000 UTC m=+1320.640520179" observedRunningTime="2026-02-27 16:45:09.548608552 +0000 UTC m=+1322.001233382" watchObservedRunningTime="2026-02-27 16:45:09.554551397 +0000 UTC m=+1322.007176247" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.594906 4814 scope.go:117] "RemoveContainer" containerID="0c9eb449398eb3654fc7114dbf4cf0886a0c7ec6574f29d26e066a62c7c64417" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.595538 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.612805 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.624320 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 16:45:09 crc kubenswrapper[4814]: E0227 16:45:09.624886 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" containerName="horizon" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.624975 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" containerName="horizon" Feb 27 16:45:09 crc kubenswrapper[4814]: E0227 16:45:09.625030 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" containerName="glance-log" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.625094 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" containerName="glance-log" Feb 27 16:45:09 crc kubenswrapper[4814]: E0227 16:45:09.625162 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa51609e-188a-4ded-962b-9f9641404afc" containerName="collect-profiles" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.625223 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa51609e-188a-4ded-962b-9f9641404afc" containerName="collect-profiles" Feb 27 16:45:09 crc kubenswrapper[4814]: E0227 16:45:09.625314 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" containerName="glance-httpd" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.625382 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" containerName="glance-httpd" Feb 27 16:45:09 crc kubenswrapper[4814]: E0227 16:45:09.625478 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" containerName="horizon-log" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.625531 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" containerName="horizon-log" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.625765 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" containerName="horizon-log" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.625831 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" containerName="glance-log" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.625887 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa51609e-188a-4ded-962b-9f9641404afc" containerName="collect-profiles" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.625949 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" containerName="glance-httpd" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.626017 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" containerName="horizon" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.627011 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.629474 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.629960 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.634184 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.761497 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4xpv\" (UniqueName: \"kubernetes.io/projected/74419238-7f7d-48d0-b138-f56913582843-kube-api-access-p4xpv\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.761544 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74419238-7f7d-48d0-b138-f56913582843-config-data\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.761573 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74419238-7f7d-48d0-b138-f56913582843-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.761618 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74419238-7f7d-48d0-b138-f56913582843-scripts\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.761815 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.761865 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74419238-7f7d-48d0-b138-f56913582843-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.761958 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74419238-7f7d-48d0-b138-f56913582843-logs\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.762093 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74419238-7f7d-48d0-b138-f56913582843-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.864226 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74419238-7f7d-48d0-b138-f56913582843-scripts\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.864321 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.864362 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74419238-7f7d-48d0-b138-f56913582843-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.864399 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74419238-7f7d-48d0-b138-f56913582843-logs\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.864464 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74419238-7f7d-48d0-b138-f56913582843-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.864488 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4xpv\" (UniqueName: \"kubernetes.io/projected/74419238-7f7d-48d0-b138-f56913582843-kube-api-access-p4xpv\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.864523 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74419238-7f7d-48d0-b138-f56913582843-config-data\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.864541 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74419238-7f7d-48d0-b138-f56913582843-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.864675 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.865373 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74419238-7f7d-48d0-b138-f56913582843-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.865492 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74419238-7f7d-48d0-b138-f56913582843-logs\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.870877 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74419238-7f7d-48d0-b138-f56913582843-scripts\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.871960 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74419238-7f7d-48d0-b138-f56913582843-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.877212 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74419238-7f7d-48d0-b138-f56913582843-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.891966 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4xpv\" (UniqueName: \"kubernetes.io/projected/74419238-7f7d-48d0-b138-f56913582843-kube-api-access-p4xpv\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.897360 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74419238-7f7d-48d0-b138-f56913582843-config-data\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.904785 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"74419238-7f7d-48d0-b138-f56913582843\") " pod="openstack/glance-default-internal-api-0" Feb 27 16:45:09 crc kubenswrapper[4814]: I0227 16:45:09.992555 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 16:45:10 crc kubenswrapper[4814]: I0227 16:45:10.499476 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f8e8e66-aaf0-43d6-b990-e046d17f2fb5" path="/var/lib/kubelet/pods/2f8e8e66-aaf0-43d6-b990-e046d17f2fb5/volumes" Feb 27 16:45:10 crc kubenswrapper[4814]: I0227 16:45:10.500692 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c9011a2-17bc-42ba-a6cf-13ec98cd75b1" path="/var/lib/kubelet/pods/6c9011a2-17bc-42ba-a6cf-13ec98cd75b1/volumes" Feb 27 16:45:10 crc kubenswrapper[4814]: W0227 16:45:10.519802 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74419238_7f7d_48d0_b138_f56913582843.slice/crio-e06dc171b90efcd190017f95993b2403c30f077e9366c91bd7d7ced4f0047494 WatchSource:0}: Error finding container e06dc171b90efcd190017f95993b2403c30f077e9366c91bd7d7ced4f0047494: Status 404 returned error can't find the container with id e06dc171b90efcd190017f95993b2403c30f077e9366c91bd7d7ced4f0047494 Feb 27 16:45:10 crc kubenswrapper[4814]: I0227 16:45:10.523598 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 16:45:10 crc kubenswrapper[4814]: I0227 16:45:10.568820 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"74419238-7f7d-48d0-b138-f56913582843","Type":"ContainerStarted","Data":"e06dc171b90efcd190017f95993b2403c30f077e9366c91bd7d7ced4f0047494"} Feb 27 16:45:10 crc kubenswrapper[4814]: I0227 16:45:10.573450 4814 generic.go:334] "Generic (PLEG): container finished" podID="0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" containerID="6f9d88326b4353e59265198023144b951d9a4aad7e4934d48508f34f04a52792" exitCode=143 Feb 27 16:45:10 crc kubenswrapper[4814]: I0227 16:45:10.574959 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31","Type":"ContainerDied","Data":"6f9d88326b4353e59265198023144b951d9a4aad7e4934d48508f34f04a52792"} Feb 27 16:45:11 crc kubenswrapper[4814]: I0227 16:45:11.587626 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"74419238-7f7d-48d0-b138-f56913582843","Type":"ContainerStarted","Data":"f17b47a36921b27f481a4c23d3ce5545ba21d25d354a3641f6db6261e4d019d3"} Feb 27 16:45:11 crc kubenswrapper[4814]: I0227 16:45:11.597637 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b665060-ed5a-4de3-91bb-606549a98922","Type":"ContainerStarted","Data":"9beda2bc38cc4afc78420df6e271c3790adc3b47e00c2b95df8d584566d5a2a6"} Feb 27 16:45:11 crc kubenswrapper[4814]: I0227 16:45:11.597807 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b665060-ed5a-4de3-91bb-606549a98922" containerName="ceilometer-central-agent" containerID="cri-o://5b43c486e1f1b2d7fc67d3724f185be74357fc71f3f713b6333cf59df95e5a60" gracePeriod=30 Feb 27 16:45:11 crc kubenswrapper[4814]: I0227 16:45:11.597931 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 16:45:11 crc kubenswrapper[4814]: I0227 16:45:11.598376 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b665060-ed5a-4de3-91bb-606549a98922" containerName="proxy-httpd" containerID="cri-o://9beda2bc38cc4afc78420df6e271c3790adc3b47e00c2b95df8d584566d5a2a6" gracePeriod=30 Feb 27 16:45:11 crc kubenswrapper[4814]: I0227 16:45:11.598405 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b665060-ed5a-4de3-91bb-606549a98922" containerName="sg-core" containerID="cri-o://fb989c51c54614706dad712ee8a0ca4d242666e2c89b1ef2f95e44d335b144ab" gracePeriod=30 Feb 27 16:45:11 crc kubenswrapper[4814]: I0227 16:45:11.598487 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b665060-ed5a-4de3-91bb-606549a98922" containerName="ceilometer-notification-agent" containerID="cri-o://0e2ee7f214497e2ad846633511393a03afe7c675c6b16de11be23b43922a86e9" gracePeriod=30 Feb 27 16:45:11 crc kubenswrapper[4814]: I0227 16:45:11.637995 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.430778439 podStartE2EDuration="12.637976991s" podCreationTimestamp="2026-02-27 16:44:59 +0000 UTC" firstStartedPulling="2026-02-27 16:45:00.378699263 +0000 UTC m=+1312.831324083" lastFinishedPulling="2026-02-27 16:45:10.585897805 +0000 UTC m=+1323.038522635" observedRunningTime="2026-02-27 16:45:11.630997573 +0000 UTC m=+1324.083622403" watchObservedRunningTime="2026-02-27 16:45:11.637976991 +0000 UTC m=+1324.090601821" Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.558938 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.152:9292/healthcheck\": read tcp 10.217.0.2:42352->10.217.0.152:9292: read: connection reset by peer" Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.559235 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.152:9292/healthcheck\": read tcp 10.217.0.2:42350->10.217.0.152:9292: read: connection reset by peer" Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.621814 4814 generic.go:334] "Generic (PLEG): container finished" podID="5b665060-ed5a-4de3-91bb-606549a98922" containerID="9beda2bc38cc4afc78420df6e271c3790adc3b47e00c2b95df8d584566d5a2a6" exitCode=0 Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.621841 4814 generic.go:334] "Generic (PLEG): container finished" podID="5b665060-ed5a-4de3-91bb-606549a98922" containerID="fb989c51c54614706dad712ee8a0ca4d242666e2c89b1ef2f95e44d335b144ab" exitCode=2 Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.621850 4814 generic.go:334] "Generic (PLEG): container finished" podID="5b665060-ed5a-4de3-91bb-606549a98922" containerID="0e2ee7f214497e2ad846633511393a03afe7c675c6b16de11be23b43922a86e9" exitCode=0 Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.621857 4814 generic.go:334] "Generic (PLEG): container finished" podID="5b665060-ed5a-4de3-91bb-606549a98922" containerID="5b43c486e1f1b2d7fc67d3724f185be74357fc71f3f713b6333cf59df95e5a60" exitCode=0 Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.621888 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b665060-ed5a-4de3-91bb-606549a98922","Type":"ContainerDied","Data":"9beda2bc38cc4afc78420df6e271c3790adc3b47e00c2b95df8d584566d5a2a6"} Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.621912 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b665060-ed5a-4de3-91bb-606549a98922","Type":"ContainerDied","Data":"fb989c51c54614706dad712ee8a0ca4d242666e2c89b1ef2f95e44d335b144ab"} Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.621921 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b665060-ed5a-4de3-91bb-606549a98922","Type":"ContainerDied","Data":"0e2ee7f214497e2ad846633511393a03afe7c675c6b16de11be23b43922a86e9"} Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.621929 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b665060-ed5a-4de3-91bb-606549a98922","Type":"ContainerDied","Data":"5b43c486e1f1b2d7fc67d3724f185be74357fc71f3f713b6333cf59df95e5a60"} Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.623298 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"74419238-7f7d-48d0-b138-f56913582843","Type":"ContainerStarted","Data":"afcef47edc1a6ede2a3aad16e3e0e4b1ef6a1ede8f52e36ed0379657270286fd"} Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.651621 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.651600934 podStartE2EDuration="3.651600934s" podCreationTimestamp="2026-02-27 16:45:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:45:12.646732362 +0000 UTC m=+1325.099357192" watchObservedRunningTime="2026-02-27 16:45:12.651600934 +0000 UTC m=+1325.104225764" Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.848940 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.920888 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-config-data\") pod \"5b665060-ed5a-4de3-91bb-606549a98922\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.921006 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-scripts\") pod \"5b665060-ed5a-4de3-91bb-606549a98922\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.921024 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-sg-core-conf-yaml\") pod \"5b665060-ed5a-4de3-91bb-606549a98922\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.921109 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdfrv\" (UniqueName: \"kubernetes.io/projected/5b665060-ed5a-4de3-91bb-606549a98922-kube-api-access-pdfrv\") pod \"5b665060-ed5a-4de3-91bb-606549a98922\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.921133 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b665060-ed5a-4de3-91bb-606549a98922-log-httpd\") pod \"5b665060-ed5a-4de3-91bb-606549a98922\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.921165 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b665060-ed5a-4de3-91bb-606549a98922-run-httpd\") pod \"5b665060-ed5a-4de3-91bb-606549a98922\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.921202 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-combined-ca-bundle\") pod \"5b665060-ed5a-4de3-91bb-606549a98922\" (UID: \"5b665060-ed5a-4de3-91bb-606549a98922\") " Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.922659 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b665060-ed5a-4de3-91bb-606549a98922-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5b665060-ed5a-4de3-91bb-606549a98922" (UID: "5b665060-ed5a-4de3-91bb-606549a98922"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.923192 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b665060-ed5a-4de3-91bb-606549a98922-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5b665060-ed5a-4de3-91bb-606549a98922" (UID: "5b665060-ed5a-4de3-91bb-606549a98922"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.929350 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-scripts" (OuterVolumeSpecName: "scripts") pod "5b665060-ed5a-4de3-91bb-606549a98922" (UID: "5b665060-ed5a-4de3-91bb-606549a98922"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.936415 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b665060-ed5a-4de3-91bb-606549a98922-kube-api-access-pdfrv" (OuterVolumeSpecName: "kube-api-access-pdfrv") pod "5b665060-ed5a-4de3-91bb-606549a98922" (UID: "5b665060-ed5a-4de3-91bb-606549a98922"). InnerVolumeSpecName "kube-api-access-pdfrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:45:12 crc kubenswrapper[4814]: I0227 16:45:12.963714 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5b665060-ed5a-4de3-91bb-606549a98922" (UID: "5b665060-ed5a-4de3-91bb-606549a98922"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.024710 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.024740 4814 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.024752 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdfrv\" (UniqueName: \"kubernetes.io/projected/5b665060-ed5a-4de3-91bb-606549a98922-kube-api-access-pdfrv\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.024761 4814 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b665060-ed5a-4de3-91bb-606549a98922-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.024768 4814 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b665060-ed5a-4de3-91bb-606549a98922-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.056935 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-config-data" (OuterVolumeSpecName: "config-data") pod "5b665060-ed5a-4de3-91bb-606549a98922" (UID: "5b665060-ed5a-4de3-91bb-606549a98922"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.059658 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b665060-ed5a-4de3-91bb-606549a98922" (UID: "5b665060-ed5a-4de3-91bb-606549a98922"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.099063 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.126070 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.126168 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-httpd-run\") pod \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.126216 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-config-data\") pod \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.126232 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-combined-ca-bundle\") pod \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.126367 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-public-tls-certs\") pod \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.126482 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-scripts\") pod \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.126497 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" (UID: "0886f8c2-6ec0-46af-8ebe-2e2cdd853d31"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.126601 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-logs\") pod \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.126637 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvmzz\" (UniqueName: \"kubernetes.io/projected/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-kube-api-access-gvmzz\") pod \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\" (UID: \"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31\") " Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.127122 4814 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.127144 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.127157 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b665060-ed5a-4de3-91bb-606549a98922-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.128193 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-logs" (OuterVolumeSpecName: "logs") pod "0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" (UID: "0886f8c2-6ec0-46af-8ebe-2e2cdd853d31"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.129433 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" (UID: "0886f8c2-6ec0-46af-8ebe-2e2cdd853d31"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.131124 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-scripts" (OuterVolumeSpecName: "scripts") pod "0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" (UID: "0886f8c2-6ec0-46af-8ebe-2e2cdd853d31"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.145958 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-kube-api-access-gvmzz" (OuterVolumeSpecName: "kube-api-access-gvmzz") pod "0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" (UID: "0886f8c2-6ec0-46af-8ebe-2e2cdd853d31"). InnerVolumeSpecName "kube-api-access-gvmzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.160900 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" (UID: "0886f8c2-6ec0-46af-8ebe-2e2cdd853d31"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.207339 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" (UID: "0886f8c2-6ec0-46af-8ebe-2e2cdd853d31"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.221502 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-config-data" (OuterVolumeSpecName: "config-data") pod "0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" (UID: "0886f8c2-6ec0-46af-8ebe-2e2cdd853d31"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.228395 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvmzz\" (UniqueName: \"kubernetes.io/projected/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-kube-api-access-gvmzz\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.228438 4814 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.228451 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.228462 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.228470 4814 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.228478 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.228486 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.252230 4814 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.330477 4814 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.638708 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b665060-ed5a-4de3-91bb-606549a98922","Type":"ContainerDied","Data":"14031101bff67b61320750949fc6a6456c79ca35cf5f608171bd0d63cc9c6b93"} Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.638780 4814 scope.go:117] "RemoveContainer" containerID="9beda2bc38cc4afc78420df6e271c3790adc3b47e00c2b95df8d584566d5a2a6" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.638855 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.642095 4814 generic.go:334] "Generic (PLEG): container finished" podID="0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" containerID="04714ba3748af9f5fe9f795ce59413df20029c8ea6f5ebf98fe3748e51f45baf" exitCode=0 Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.642169 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.642186 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31","Type":"ContainerDied","Data":"04714ba3748af9f5fe9f795ce59413df20029c8ea6f5ebf98fe3748e51f45baf"} Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.642245 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0886f8c2-6ec0-46af-8ebe-2e2cdd853d31","Type":"ContainerDied","Data":"93513a28d90cb7ec06a4af69840230c2530d021ffcb4ad375e1bd50fa030547f"} Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.673563 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.680198 4814 scope.go:117] "RemoveContainer" containerID="fb989c51c54614706dad712ee8a0ca4d242666e2c89b1ef2f95e44d335b144ab" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.680491 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.704168 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.707957 4814 scope.go:117] "RemoveContainer" containerID="0e2ee7f214497e2ad846633511393a03afe7c675c6b16de11be23b43922a86e9" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.720673 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.728874 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 16:45:13 crc kubenswrapper[4814]: E0227 16:45:13.729195 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" containerName="glance-log" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.729211 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" containerName="glance-log" Feb 27 16:45:13 crc kubenswrapper[4814]: E0227 16:45:13.729229 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" containerName="glance-httpd" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.729235 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" containerName="glance-httpd" Feb 27 16:45:13 crc kubenswrapper[4814]: E0227 16:45:13.729245 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b665060-ed5a-4de3-91bb-606549a98922" containerName="ceilometer-central-agent" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.729269 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b665060-ed5a-4de3-91bb-606549a98922" containerName="ceilometer-central-agent" Feb 27 16:45:13 crc kubenswrapper[4814]: E0227 16:45:13.729286 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b665060-ed5a-4de3-91bb-606549a98922" containerName="ceilometer-notification-agent" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.729292 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b665060-ed5a-4de3-91bb-606549a98922" containerName="ceilometer-notification-agent" Feb 27 16:45:13 crc kubenswrapper[4814]: E0227 16:45:13.729303 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b665060-ed5a-4de3-91bb-606549a98922" containerName="sg-core" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.729309 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b665060-ed5a-4de3-91bb-606549a98922" containerName="sg-core" Feb 27 16:45:13 crc kubenswrapper[4814]: E0227 16:45:13.729328 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b665060-ed5a-4de3-91bb-606549a98922" containerName="proxy-httpd" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.729334 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b665060-ed5a-4de3-91bb-606549a98922" containerName="proxy-httpd" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.729488 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" containerName="glance-httpd" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.729501 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b665060-ed5a-4de3-91bb-606549a98922" containerName="ceilometer-notification-agent" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.729513 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" containerName="glance-log" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.729521 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b665060-ed5a-4de3-91bb-606549a98922" containerName="sg-core" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.729532 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b665060-ed5a-4de3-91bb-606549a98922" containerName="ceilometer-central-agent" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.729540 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b665060-ed5a-4de3-91bb-606549a98922" containerName="proxy-httpd" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.730394 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.735239 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.735513 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.737751 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.740411 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.742602 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.743088 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.745034 4814 scope.go:117] "RemoveContainer" containerID="5b43c486e1f1b2d7fc67d3724f185be74357fc71f3f713b6333cf59df95e5a60" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.752394 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.759548 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.837173 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f71993c-f340-475c-a9be-0dcd142fdfd2-log-httpd\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.837208 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhjkn\" (UniqueName: \"kubernetes.io/projected/9f71993c-f340-475c-a9be-0dcd142fdfd2-kube-api-access-dhjkn\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.837239 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5skk\" (UniqueName: \"kubernetes.io/projected/1e537a87-2e08-4521-bfa5-aeecdc14159d-kube-api-access-t5skk\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.837297 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-scripts\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.837327 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e537a87-2e08-4521-bfa5-aeecdc14159d-logs\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.837356 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-config-data\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.837393 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e537a87-2e08-4521-bfa5-aeecdc14159d-scripts\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.837409 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.837436 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e537a87-2e08-4521-bfa5-aeecdc14159d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.837456 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.837495 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f71993c-f340-475c-a9be-0dcd142fdfd2-run-httpd\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.837515 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.837535 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1e537a87-2e08-4521-bfa5-aeecdc14159d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.837552 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e537a87-2e08-4521-bfa5-aeecdc14159d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.837586 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e537a87-2e08-4521-bfa5-aeecdc14159d-config-data\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.861917 4814 scope.go:117] "RemoveContainer" containerID="04714ba3748af9f5fe9f795ce59413df20029c8ea6f5ebf98fe3748e51f45baf" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.887360 4814 scope.go:117] "RemoveContainer" containerID="6f9d88326b4353e59265198023144b951d9a4aad7e4934d48508f34f04a52792" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.911272 4814 scope.go:117] "RemoveContainer" containerID="04714ba3748af9f5fe9f795ce59413df20029c8ea6f5ebf98fe3748e51f45baf" Feb 27 16:45:13 crc kubenswrapper[4814]: E0227 16:45:13.911679 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04714ba3748af9f5fe9f795ce59413df20029c8ea6f5ebf98fe3748e51f45baf\": container with ID starting with 04714ba3748af9f5fe9f795ce59413df20029c8ea6f5ebf98fe3748e51f45baf not found: ID does not exist" containerID="04714ba3748af9f5fe9f795ce59413df20029c8ea6f5ebf98fe3748e51f45baf" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.911723 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04714ba3748af9f5fe9f795ce59413df20029c8ea6f5ebf98fe3748e51f45baf"} err="failed to get container status \"04714ba3748af9f5fe9f795ce59413df20029c8ea6f5ebf98fe3748e51f45baf\": rpc error: code = NotFound desc = could not find container \"04714ba3748af9f5fe9f795ce59413df20029c8ea6f5ebf98fe3748e51f45baf\": container with ID starting with 04714ba3748af9f5fe9f795ce59413df20029c8ea6f5ebf98fe3748e51f45baf not found: ID does not exist" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.911748 4814 scope.go:117] "RemoveContainer" containerID="6f9d88326b4353e59265198023144b951d9a4aad7e4934d48508f34f04a52792" Feb 27 16:45:13 crc kubenswrapper[4814]: E0227 16:45:13.912098 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f9d88326b4353e59265198023144b951d9a4aad7e4934d48508f34f04a52792\": container with ID starting with 6f9d88326b4353e59265198023144b951d9a4aad7e4934d48508f34f04a52792 not found: ID does not exist" containerID="6f9d88326b4353e59265198023144b951d9a4aad7e4934d48508f34f04a52792" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.912129 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f9d88326b4353e59265198023144b951d9a4aad7e4934d48508f34f04a52792"} err="failed to get container status \"6f9d88326b4353e59265198023144b951d9a4aad7e4934d48508f34f04a52792\": rpc error: code = NotFound desc = could not find container \"6f9d88326b4353e59265198023144b951d9a4aad7e4934d48508f34f04a52792\": container with ID starting with 6f9d88326b4353e59265198023144b951d9a4aad7e4934d48508f34f04a52792 not found: ID does not exist" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.939692 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-config-data\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.939763 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e537a87-2e08-4521-bfa5-aeecdc14159d-scripts\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.939784 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.939799 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e537a87-2e08-4521-bfa5-aeecdc14159d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.939822 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.939867 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f71993c-f340-475c-a9be-0dcd142fdfd2-run-httpd\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.939888 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.939914 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1e537a87-2e08-4521-bfa5-aeecdc14159d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.939932 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e537a87-2e08-4521-bfa5-aeecdc14159d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.939968 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e537a87-2e08-4521-bfa5-aeecdc14159d-config-data\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.940000 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f71993c-f340-475c-a9be-0dcd142fdfd2-log-httpd\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.940016 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhjkn\" (UniqueName: \"kubernetes.io/projected/9f71993c-f340-475c-a9be-0dcd142fdfd2-kube-api-access-dhjkn\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.940041 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5skk\" (UniqueName: \"kubernetes.io/projected/1e537a87-2e08-4521-bfa5-aeecdc14159d-kube-api-access-t5skk\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.940043 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.940057 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-scripts\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.940074 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e537a87-2e08-4521-bfa5-aeecdc14159d-logs\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.940567 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f71993c-f340-475c-a9be-0dcd142fdfd2-log-httpd\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.940643 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f71993c-f340-475c-a9be-0dcd142fdfd2-run-httpd\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.941593 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e537a87-2e08-4521-bfa5-aeecdc14159d-logs\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.942998 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1e537a87-2e08-4521-bfa5-aeecdc14159d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.944009 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e537a87-2e08-4521-bfa5-aeecdc14159d-scripts\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.944698 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-scripts\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.944835 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e537a87-2e08-4521-bfa5-aeecdc14159d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.949844 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e537a87-2e08-4521-bfa5-aeecdc14159d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.953931 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.955480 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.957227 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e537a87-2e08-4521-bfa5-aeecdc14159d-config-data\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.957487 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhjkn\" (UniqueName: \"kubernetes.io/projected/9f71993c-f340-475c-a9be-0dcd142fdfd2-kube-api-access-dhjkn\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.959758 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5skk\" (UniqueName: \"kubernetes.io/projected/1e537a87-2e08-4521-bfa5-aeecdc14159d-kube-api-access-t5skk\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.986230 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-config-data\") pod \"ceilometer-0\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " pod="openstack/ceilometer-0" Feb 27 16:45:13 crc kubenswrapper[4814]: I0227 16:45:13.995292 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"1e537a87-2e08-4521-bfa5-aeecdc14159d\") " pod="openstack/glance-default-external-api-0" Feb 27 16:45:14 crc kubenswrapper[4814]: I0227 16:45:14.073814 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 16:45:14 crc kubenswrapper[4814]: I0227 16:45:14.135967 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:45:14 crc kubenswrapper[4814]: I0227 16:45:14.498941 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0886f8c2-6ec0-46af-8ebe-2e2cdd853d31" path="/var/lib/kubelet/pods/0886f8c2-6ec0-46af-8ebe-2e2cdd853d31/volumes" Feb 27 16:45:14 crc kubenswrapper[4814]: I0227 16:45:14.499563 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b665060-ed5a-4de3-91bb-606549a98922" path="/var/lib/kubelet/pods/5b665060-ed5a-4de3-91bb-606549a98922/volumes" Feb 27 16:45:14 crc kubenswrapper[4814]: I0227 16:45:14.724324 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:45:14 crc kubenswrapper[4814]: I0227 16:45:14.763235 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 16:45:15 crc kubenswrapper[4814]: I0227 16:45:15.664075 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f71993c-f340-475c-a9be-0dcd142fdfd2","Type":"ContainerStarted","Data":"03885dc6382f4861afc9978da099f54d7e609651bd191f79e46042814ea1e796"} Feb 27 16:45:15 crc kubenswrapper[4814]: I0227 16:45:15.664373 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f71993c-f340-475c-a9be-0dcd142fdfd2","Type":"ContainerStarted","Data":"3776913b7188a16ef1a4b51611f9617689d5ec913eccc539361c21e936b8d910"} Feb 27 16:45:15 crc kubenswrapper[4814]: I0227 16:45:15.665656 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1e537a87-2e08-4521-bfa5-aeecdc14159d","Type":"ContainerStarted","Data":"4a04beba8044de3324673be71d247c91e5ffef5e880e54eac97f2d134373499a"} Feb 27 16:45:15 crc kubenswrapper[4814]: I0227 16:45:15.665690 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1e537a87-2e08-4521-bfa5-aeecdc14159d","Type":"ContainerStarted","Data":"39f128bd153223b0c2c3fb4edd6d0b970712436d220c6d6108828f4f0bb6159e"} Feb 27 16:45:16 crc kubenswrapper[4814]: I0227 16:45:16.685208 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1e537a87-2e08-4521-bfa5-aeecdc14159d","Type":"ContainerStarted","Data":"2feb201ba4301bb2a78bd633fecf6609287af5d3ff4e7f0b5f1465a8b4cfdcae"} Feb 27 16:45:16 crc kubenswrapper[4814]: I0227 16:45:16.690192 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f71993c-f340-475c-a9be-0dcd142fdfd2","Type":"ContainerStarted","Data":"9fb6c810b4d67c671f9213672bb6b6e59b6373ec325597e0d2b94b636234bdde"} Feb 27 16:45:16 crc kubenswrapper[4814]: I0227 16:45:16.719874 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.719858768 podStartE2EDuration="3.719858768s" podCreationTimestamp="2026-02-27 16:45:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:45:16.713512639 +0000 UTC m=+1329.166137469" watchObservedRunningTime="2026-02-27 16:45:16.719858768 +0000 UTC m=+1329.172483598" Feb 27 16:45:17 crc kubenswrapper[4814]: I0227 16:45:17.702745 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f71993c-f340-475c-a9be-0dcd142fdfd2","Type":"ContainerStarted","Data":"ecf99d3800e43891e17b5b83917e59cfd7606fad68fc796fe5d2ce2ee3185206"} Feb 27 16:45:18 crc kubenswrapper[4814]: I0227 16:45:18.717401 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f71993c-f340-475c-a9be-0dcd142fdfd2","Type":"ContainerStarted","Data":"e885e708ab2e43f55848a6c02956ddaa3f07e0c066b53124a2311901ef215ec7"} Feb 27 16:45:18 crc kubenswrapper[4814]: I0227 16:45:18.718328 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 16:45:18 crc kubenswrapper[4814]: I0227 16:45:18.747719 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.104635354 podStartE2EDuration="5.747695803s" podCreationTimestamp="2026-02-27 16:45:13 +0000 UTC" firstStartedPulling="2026-02-27 16:45:14.696459992 +0000 UTC m=+1327.149084822" lastFinishedPulling="2026-02-27 16:45:18.339520441 +0000 UTC m=+1330.792145271" observedRunningTime="2026-02-27 16:45:18.738477924 +0000 UTC m=+1331.191102754" watchObservedRunningTime="2026-02-27 16:45:18.747695803 +0000 UTC m=+1331.200320643" Feb 27 16:45:19 crc kubenswrapper[4814]: E0227 16:45:19.164476 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa51609e_188a_4ded_962b_9f9641404afc.slice/crio-eb7a32b82b3fba9fb45c72f26ffdb9697916b2788f479c2eb70fe921070b3abe\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa51609e_188a_4ded_962b_9f9641404afc.slice\": RecentStats: unable to find data in memory cache]" Feb 27 16:45:19 crc kubenswrapper[4814]: I0227 16:45:19.233681 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:45:19 crc kubenswrapper[4814]: I0227 16:45:19.993651 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 27 16:45:19 crc kubenswrapper[4814]: I0227 16:45:19.994508 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 27 16:45:20 crc kubenswrapper[4814]: I0227 16:45:20.033136 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 27 16:45:20 crc kubenswrapper[4814]: I0227 16:45:20.044564 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 27 16:45:20 crc kubenswrapper[4814]: I0227 16:45:20.737741 4814 generic.go:334] "Generic (PLEG): container finished" podID="827ecd26-d313-4fb7-86bb-03e6e777f5a7" containerID="5f2cc4f12adcccdec8db26f18084d239bc8397263cda448b3eb475afe5553635" exitCode=0 Feb 27 16:45:20 crc kubenswrapper[4814]: I0227 16:45:20.737980 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jjmg8" event={"ID":"827ecd26-d313-4fb7-86bb-03e6e777f5a7","Type":"ContainerDied","Data":"5f2cc4f12adcccdec8db26f18084d239bc8397263cda448b3eb475afe5553635"} Feb 27 16:45:20 crc kubenswrapper[4814]: I0227 16:45:20.739447 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 27 16:45:20 crc kubenswrapper[4814]: I0227 16:45:20.739774 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 27 16:45:20 crc kubenswrapper[4814]: I0227 16:45:20.739811 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerName="ceilometer-central-agent" containerID="cri-o://03885dc6382f4861afc9978da099f54d7e609651bd191f79e46042814ea1e796" gracePeriod=30 Feb 27 16:45:20 crc kubenswrapper[4814]: I0227 16:45:20.739848 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerName="proxy-httpd" containerID="cri-o://e885e708ab2e43f55848a6c02956ddaa3f07e0c066b53124a2311901ef215ec7" gracePeriod=30 Feb 27 16:45:20 crc kubenswrapper[4814]: I0227 16:45:20.740009 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerName="ceilometer-notification-agent" containerID="cri-o://9fb6c810b4d67c671f9213672bb6b6e59b6373ec325597e0d2b94b636234bdde" gracePeriod=30 Feb 27 16:45:20 crc kubenswrapper[4814]: I0227 16:45:20.740091 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerName="sg-core" containerID="cri-o://ecf99d3800e43891e17b5b83917e59cfd7606fad68fc796fe5d2ce2ee3185206" gracePeriod=30 Feb 27 16:45:21 crc kubenswrapper[4814]: I0227 16:45:21.749874 4814 generic.go:334] "Generic (PLEG): container finished" podID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerID="e885e708ab2e43f55848a6c02956ddaa3f07e0c066b53124a2311901ef215ec7" exitCode=0 Feb 27 16:45:21 crc kubenswrapper[4814]: I0227 16:45:21.750120 4814 generic.go:334] "Generic (PLEG): container finished" podID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerID="ecf99d3800e43891e17b5b83917e59cfd7606fad68fc796fe5d2ce2ee3185206" exitCode=2 Feb 27 16:45:21 crc kubenswrapper[4814]: I0227 16:45:21.750128 4814 generic.go:334] "Generic (PLEG): container finished" podID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerID="9fb6c810b4d67c671f9213672bb6b6e59b6373ec325597e0d2b94b636234bdde" exitCode=0 Feb 27 16:45:21 crc kubenswrapper[4814]: I0227 16:45:21.749964 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f71993c-f340-475c-a9be-0dcd142fdfd2","Type":"ContainerDied","Data":"e885e708ab2e43f55848a6c02956ddaa3f07e0c066b53124a2311901ef215ec7"} Feb 27 16:45:21 crc kubenswrapper[4814]: I0227 16:45:21.750197 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f71993c-f340-475c-a9be-0dcd142fdfd2","Type":"ContainerDied","Data":"ecf99d3800e43891e17b5b83917e59cfd7606fad68fc796fe5d2ce2ee3185206"} Feb 27 16:45:21 crc kubenswrapper[4814]: I0227 16:45:21.750222 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f71993c-f340-475c-a9be-0dcd142fdfd2","Type":"ContainerDied","Data":"9fb6c810b4d67c671f9213672bb6b6e59b6373ec325597e0d2b94b636234bdde"} Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.123097 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jjmg8" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.139740 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-combined-ca-bundle\") pod \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\" (UID: \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\") " Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.139793 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-config-data\") pod \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\" (UID: \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\") " Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.139894 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-scripts\") pod \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\" (UID: \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\") " Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.140012 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnm8l\" (UniqueName: \"kubernetes.io/projected/827ecd26-d313-4fb7-86bb-03e6e777f5a7-kube-api-access-jnm8l\") pod \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\" (UID: \"827ecd26-d313-4fb7-86bb-03e6e777f5a7\") " Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.147547 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/827ecd26-d313-4fb7-86bb-03e6e777f5a7-kube-api-access-jnm8l" (OuterVolumeSpecName: "kube-api-access-jnm8l") pod "827ecd26-d313-4fb7-86bb-03e6e777f5a7" (UID: "827ecd26-d313-4fb7-86bb-03e6e777f5a7"). InnerVolumeSpecName "kube-api-access-jnm8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.152125 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-scripts" (OuterVolumeSpecName: "scripts") pod "827ecd26-d313-4fb7-86bb-03e6e777f5a7" (UID: "827ecd26-d313-4fb7-86bb-03e6e777f5a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.177377 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "827ecd26-d313-4fb7-86bb-03e6e777f5a7" (UID: "827ecd26-d313-4fb7-86bb-03e6e777f5a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.192414 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-config-data" (OuterVolumeSpecName: "config-data") pod "827ecd26-d313-4fb7-86bb-03e6e777f5a7" (UID: "827ecd26-d313-4fb7-86bb-03e6e777f5a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.261916 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.261957 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnm8l\" (UniqueName: \"kubernetes.io/projected/827ecd26-d313-4fb7-86bb-03e6e777f5a7-kube-api-access-jnm8l\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.261973 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.261991 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/827ecd26-d313-4fb7-86bb-03e6e777f5a7-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.761402 4814 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.762382 4814 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.762677 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jjmg8" event={"ID":"827ecd26-d313-4fb7-86bb-03e6e777f5a7","Type":"ContainerDied","Data":"5eb131b727f1989859a6f39d6a1f16ad3149c04f1cbeb255d326d107afad74c6"} Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.762785 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5eb131b727f1989859a6f39d6a1f16ad3149c04f1cbeb255d326d107afad74c6" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.765045 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jjmg8" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.890333 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 27 16:45:22 crc kubenswrapper[4814]: E0227 16:45:22.890782 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="827ecd26-d313-4fb7-86bb-03e6e777f5a7" containerName="nova-cell0-conductor-db-sync" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.890796 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="827ecd26-d313-4fb7-86bb-03e6e777f5a7" containerName="nova-cell0-conductor-db-sync" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.890948 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="827ecd26-d313-4fb7-86bb-03e6e777f5a7" containerName="nova-cell0-conductor-db-sync" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.891562 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.902610 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.902611 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.902734 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.902846 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-x2grh" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.907649 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.910664 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.915093 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.978121 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np8vp\" (UniqueName: \"kubernetes.io/projected/f3152b8d-6ee7-4107-9205-b3535102b3cb-kube-api-access-np8vp\") pod \"nova-cell0-conductor-0\" (UID: \"f3152b8d-6ee7-4107-9205-b3535102b3cb\") " pod="openstack/nova-cell0-conductor-0" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.978291 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3152b8d-6ee7-4107-9205-b3535102b3cb-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f3152b8d-6ee7-4107-9205-b3535102b3cb\") " pod="openstack/nova-cell0-conductor-0" Feb 27 16:45:22 crc kubenswrapper[4814]: I0227 16:45:22.978371 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3152b8d-6ee7-4107-9205-b3535102b3cb-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f3152b8d-6ee7-4107-9205-b3535102b3cb\") " pod="openstack/nova-cell0-conductor-0" Feb 27 16:45:23 crc kubenswrapper[4814]: I0227 16:45:23.079611 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3152b8d-6ee7-4107-9205-b3535102b3cb-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f3152b8d-6ee7-4107-9205-b3535102b3cb\") " pod="openstack/nova-cell0-conductor-0" Feb 27 16:45:23 crc kubenswrapper[4814]: I0227 16:45:23.079730 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np8vp\" (UniqueName: \"kubernetes.io/projected/f3152b8d-6ee7-4107-9205-b3535102b3cb-kube-api-access-np8vp\") pod \"nova-cell0-conductor-0\" (UID: \"f3152b8d-6ee7-4107-9205-b3535102b3cb\") " pod="openstack/nova-cell0-conductor-0" Feb 27 16:45:23 crc kubenswrapper[4814]: I0227 16:45:23.079779 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3152b8d-6ee7-4107-9205-b3535102b3cb-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f3152b8d-6ee7-4107-9205-b3535102b3cb\") " pod="openstack/nova-cell0-conductor-0" Feb 27 16:45:23 crc kubenswrapper[4814]: I0227 16:45:23.084538 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3152b8d-6ee7-4107-9205-b3535102b3cb-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f3152b8d-6ee7-4107-9205-b3535102b3cb\") " pod="openstack/nova-cell0-conductor-0" Feb 27 16:45:23 crc kubenswrapper[4814]: I0227 16:45:23.084943 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3152b8d-6ee7-4107-9205-b3535102b3cb-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f3152b8d-6ee7-4107-9205-b3535102b3cb\") " pod="openstack/nova-cell0-conductor-0" Feb 27 16:45:23 crc kubenswrapper[4814]: I0227 16:45:23.099025 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np8vp\" (UniqueName: \"kubernetes.io/projected/f3152b8d-6ee7-4107-9205-b3535102b3cb-kube-api-access-np8vp\") pod \"nova-cell0-conductor-0\" (UID: \"f3152b8d-6ee7-4107-9205-b3535102b3cb\") " pod="openstack/nova-cell0-conductor-0" Feb 27 16:45:23 crc kubenswrapper[4814]: I0227 16:45:23.230493 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 27 16:45:23 crc kubenswrapper[4814]: I0227 16:45:23.694503 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 27 16:45:23 crc kubenswrapper[4814]: I0227 16:45:23.779783 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f3152b8d-6ee7-4107-9205-b3535102b3cb","Type":"ContainerStarted","Data":"b8c4438f6fc582a71d5c2f73d45d18bf38dbe138042c225f31f4cf84b6ff1f73"} Feb 27 16:45:24 crc kubenswrapper[4814]: I0227 16:45:24.074424 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 27 16:45:24 crc kubenswrapper[4814]: I0227 16:45:24.074457 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 27 16:45:24 crc kubenswrapper[4814]: I0227 16:45:24.118742 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 27 16:45:24 crc kubenswrapper[4814]: I0227 16:45:24.124288 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 27 16:45:24 crc kubenswrapper[4814]: I0227 16:45:24.794994 4814 generic.go:334] "Generic (PLEG): container finished" podID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerID="03885dc6382f4861afc9978da099f54d7e609651bd191f79e46042814ea1e796" exitCode=0 Feb 27 16:45:24 crc kubenswrapper[4814]: I0227 16:45:24.795233 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f71993c-f340-475c-a9be-0dcd142fdfd2","Type":"ContainerDied","Data":"03885dc6382f4861afc9978da099f54d7e609651bd191f79e46042814ea1e796"} Feb 27 16:45:24 crc kubenswrapper[4814]: I0227 16:45:24.797510 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f3152b8d-6ee7-4107-9205-b3535102b3cb","Type":"ContainerStarted","Data":"e745aad141d4beb9bb3c2cb6010659a46c5657d0c08b35cff1d2772ecbbbf8b3"} Feb 27 16:45:24 crc kubenswrapper[4814]: I0227 16:45:24.799007 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 27 16:45:24 crc kubenswrapper[4814]: I0227 16:45:24.799077 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 27 16:45:24 crc kubenswrapper[4814]: I0227 16:45:24.825820 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.825795228 podStartE2EDuration="2.825795228s" podCreationTimestamp="2026-02-27 16:45:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:45:24.824666682 +0000 UTC m=+1337.277291542" watchObservedRunningTime="2026-02-27 16:45:24.825795228 +0000 UTC m=+1337.278420088" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.440090 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.533103 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-config-data\") pod \"9f71993c-f340-475c-a9be-0dcd142fdfd2\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.533243 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f71993c-f340-475c-a9be-0dcd142fdfd2-log-httpd\") pod \"9f71993c-f340-475c-a9be-0dcd142fdfd2\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.533344 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f71993c-f340-475c-a9be-0dcd142fdfd2-run-httpd\") pod \"9f71993c-f340-475c-a9be-0dcd142fdfd2\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.533376 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhjkn\" (UniqueName: \"kubernetes.io/projected/9f71993c-f340-475c-a9be-0dcd142fdfd2-kube-api-access-dhjkn\") pod \"9f71993c-f340-475c-a9be-0dcd142fdfd2\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.533439 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-sg-core-conf-yaml\") pod \"9f71993c-f340-475c-a9be-0dcd142fdfd2\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.533718 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-combined-ca-bundle\") pod \"9f71993c-f340-475c-a9be-0dcd142fdfd2\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.533853 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-scripts\") pod \"9f71993c-f340-475c-a9be-0dcd142fdfd2\" (UID: \"9f71993c-f340-475c-a9be-0dcd142fdfd2\") " Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.534882 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f71993c-f340-475c-a9be-0dcd142fdfd2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9f71993c-f340-475c-a9be-0dcd142fdfd2" (UID: "9f71993c-f340-475c-a9be-0dcd142fdfd2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.535058 4814 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f71993c-f340-475c-a9be-0dcd142fdfd2-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.535397 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f71993c-f340-475c-a9be-0dcd142fdfd2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9f71993c-f340-475c-a9be-0dcd142fdfd2" (UID: "9f71993c-f340-475c-a9be-0dcd142fdfd2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.542515 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f71993c-f340-475c-a9be-0dcd142fdfd2-kube-api-access-dhjkn" (OuterVolumeSpecName: "kube-api-access-dhjkn") pod "9f71993c-f340-475c-a9be-0dcd142fdfd2" (UID: "9f71993c-f340-475c-a9be-0dcd142fdfd2"). InnerVolumeSpecName "kube-api-access-dhjkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.556585 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-scripts" (OuterVolumeSpecName: "scripts") pod "9f71993c-f340-475c-a9be-0dcd142fdfd2" (UID: "9f71993c-f340-475c-a9be-0dcd142fdfd2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.560679 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9f71993c-f340-475c-a9be-0dcd142fdfd2" (UID: "9f71993c-f340-475c-a9be-0dcd142fdfd2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.618797 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f71993c-f340-475c-a9be-0dcd142fdfd2" (UID: "9f71993c-f340-475c-a9be-0dcd142fdfd2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.636980 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhjkn\" (UniqueName: \"kubernetes.io/projected/9f71993c-f340-475c-a9be-0dcd142fdfd2-kube-api-access-dhjkn\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.637024 4814 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.637040 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.637058 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.637073 4814 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f71993c-f340-475c-a9be-0dcd142fdfd2-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.641458 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-config-data" (OuterVolumeSpecName: "config-data") pod "9f71993c-f340-475c-a9be-0dcd142fdfd2" (UID: "9f71993c-f340-475c-a9be-0dcd142fdfd2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.738740 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f71993c-f340-475c-a9be-0dcd142fdfd2-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.814174 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f71993c-f340-475c-a9be-0dcd142fdfd2","Type":"ContainerDied","Data":"3776913b7188a16ef1a4b51611f9617689d5ec913eccc539361c21e936b8d910"} Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.814287 4814 scope.go:117] "RemoveContainer" containerID="e885e708ab2e43f55848a6c02956ddaa3f07e0c066b53124a2311901ef215ec7" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.814698 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.815735 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.840437 4814 scope.go:117] "RemoveContainer" containerID="ecf99d3800e43891e17b5b83917e59cfd7606fad68fc796fe5d2ce2ee3185206" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.868299 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.878536 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.900487 4814 scope.go:117] "RemoveContainer" containerID="9fb6c810b4d67c671f9213672bb6b6e59b6373ec325597e0d2b94b636234bdde" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.921948 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:45:25 crc kubenswrapper[4814]: E0227 16:45:25.922615 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerName="ceilometer-notification-agent" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.922647 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerName="ceilometer-notification-agent" Feb 27 16:45:25 crc kubenswrapper[4814]: E0227 16:45:25.922673 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerName="sg-core" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.922687 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerName="sg-core" Feb 27 16:45:25 crc kubenswrapper[4814]: E0227 16:45:25.922720 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerName="proxy-httpd" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.922733 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerName="proxy-httpd" Feb 27 16:45:25 crc kubenswrapper[4814]: E0227 16:45:25.922758 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerName="ceilometer-central-agent" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.922770 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerName="ceilometer-central-agent" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.923101 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerName="ceilometer-central-agent" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.923138 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerName="sg-core" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.923168 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerName="proxy-httpd" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.923190 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" containerName="ceilometer-notification-agent" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.926009 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.935669 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.935990 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.966883 4814 scope.go:117] "RemoveContainer" containerID="03885dc6382f4861afc9978da099f54d7e609651bd191f79e46042814ea1e796" Feb 27 16:45:25 crc kubenswrapper[4814]: I0227 16:45:25.971343 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.045483 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-scripts\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.045541 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-config-data\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.045593 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dd4dce7-515e-41bf-9fef-868960c262d7-log-httpd\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.045635 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.045669 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dd4dce7-515e-41bf-9fef-868960c262d7-run-httpd\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.045717 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.045904 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8sww\" (UniqueName: \"kubernetes.io/projected/7dd4dce7-515e-41bf-9fef-868960c262d7-kube-api-access-k8sww\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.147938 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8sww\" (UniqueName: \"kubernetes.io/projected/7dd4dce7-515e-41bf-9fef-868960c262d7-kube-api-access-k8sww\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.148599 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-scripts\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.148808 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-config-data\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.148944 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dd4dce7-515e-41bf-9fef-868960c262d7-log-httpd\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.149070 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.149187 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dd4dce7-515e-41bf-9fef-868960c262d7-run-httpd\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.149381 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.150704 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dd4dce7-515e-41bf-9fef-868960c262d7-run-httpd\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.152311 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dd4dce7-515e-41bf-9fef-868960c262d7-log-httpd\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.154396 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-config-data\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.155725 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.156455 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.159908 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-scripts\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.178084 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8sww\" (UniqueName: \"kubernetes.io/projected/7dd4dce7-515e-41bf-9fef-868960c262d7-kube-api-access-k8sww\") pod \"ceilometer-0\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.254082 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.509683 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f71993c-f340-475c-a9be-0dcd142fdfd2" path="/var/lib/kubelet/pods/9f71993c-f340-475c-a9be-0dcd142fdfd2/volumes" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.572618 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.590019 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:45:26 crc kubenswrapper[4814]: W0227 16:45:26.593544 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dd4dce7_515e_41bf_9fef_868960c262d7.slice/crio-13935b723ba812af0c8e8b8842a40bebb8304d6afeaf6d01dcd2b2cf4abb8531 WatchSource:0}: Error finding container 13935b723ba812af0c8e8b8842a40bebb8304d6afeaf6d01dcd2b2cf4abb8531: Status 404 returned error can't find the container with id 13935b723ba812af0c8e8b8842a40bebb8304d6afeaf6d01dcd2b2cf4abb8531 Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.713833 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 27 16:45:26 crc kubenswrapper[4814]: I0227 16:45:26.825358 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dd4dce7-515e-41bf-9fef-868960c262d7","Type":"ContainerStarted","Data":"13935b723ba812af0c8e8b8842a40bebb8304d6afeaf6d01dcd2b2cf4abb8531"} Feb 27 16:45:27 crc kubenswrapper[4814]: I0227 16:45:27.838699 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dd4dce7-515e-41bf-9fef-868960c262d7","Type":"ContainerStarted","Data":"f2a1a11a333a71952c7354ceba5d5534bfd7cd11d4a5fd92841903e1b80046d6"} Feb 27 16:45:28 crc kubenswrapper[4814]: I0227 16:45:28.853272 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dd4dce7-515e-41bf-9fef-868960c262d7","Type":"ContainerStarted","Data":"3c93381780739912b02366f3b4a4ed1ec7a7fa4b551cea3ed06fa6c3d954d8d3"} Feb 27 16:45:28 crc kubenswrapper[4814]: I0227 16:45:28.853663 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dd4dce7-515e-41bf-9fef-868960c262d7","Type":"ContainerStarted","Data":"5d4a58b5c3973b8d8eb818fa89598227387781a513be3d96579efd1b35d9fd0b"} Feb 27 16:45:29 crc kubenswrapper[4814]: E0227 16:45:29.366595 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa51609e_188a_4ded_962b_9f9641404afc.slice/crio-eb7a32b82b3fba9fb45c72f26ffdb9697916b2788f479c2eb70fe921070b3abe\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa51609e_188a_4ded_962b_9f9641404afc.slice\": RecentStats: unable to find data in memory cache]" Feb 27 16:45:31 crc kubenswrapper[4814]: I0227 16:45:31.904652 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dd4dce7-515e-41bf-9fef-868960c262d7","Type":"ContainerStarted","Data":"fbe05b43134f6570f2e5fcb9e50bcaacf6cb9dac269d6014816dbe35e87cd8c1"} Feb 27 16:45:31 crc kubenswrapper[4814]: I0227 16:45:31.905338 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 16:45:31 crc kubenswrapper[4814]: I0227 16:45:31.943782 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.916695262 podStartE2EDuration="6.943756327s" podCreationTimestamp="2026-02-27 16:45:25 +0000 UTC" firstStartedPulling="2026-02-27 16:45:26.598896922 +0000 UTC m=+1339.051521752" lastFinishedPulling="2026-02-27 16:45:30.625957957 +0000 UTC m=+1343.078582817" observedRunningTime="2026-02-27 16:45:31.93458422 +0000 UTC m=+1344.387209090" watchObservedRunningTime="2026-02-27 16:45:31.943756327 +0000 UTC m=+1344.396381197" Feb 27 16:45:33 crc kubenswrapper[4814]: I0227 16:45:33.282515 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 27 16:45:33 crc kubenswrapper[4814]: I0227 16:45:33.861852 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-52n2r"] Feb 27 16:45:33 crc kubenswrapper[4814]: I0227 16:45:33.863588 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-52n2r" Feb 27 16:45:33 crc kubenswrapper[4814]: I0227 16:45:33.870095 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 27 16:45:33 crc kubenswrapper[4814]: I0227 16:45:33.870496 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 27 16:45:33 crc kubenswrapper[4814]: I0227 16:45:33.882176 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-52n2r"] Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.008560 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc8d7\" (UniqueName: \"kubernetes.io/projected/66b6902c-c560-43ba-8bf6-b162da3931df-kube-api-access-zc8d7\") pod \"nova-cell0-cell-mapping-52n2r\" (UID: \"66b6902c-c560-43ba-8bf6-b162da3931df\") " pod="openstack/nova-cell0-cell-mapping-52n2r" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.009008 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-scripts\") pod \"nova-cell0-cell-mapping-52n2r\" (UID: \"66b6902c-c560-43ba-8bf6-b162da3931df\") " pod="openstack/nova-cell0-cell-mapping-52n2r" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.009092 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-config-data\") pod \"nova-cell0-cell-mapping-52n2r\" (UID: \"66b6902c-c560-43ba-8bf6-b162da3931df\") " pod="openstack/nova-cell0-cell-mapping-52n2r" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.009129 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-52n2r\" (UID: \"66b6902c-c560-43ba-8bf6-b162da3931df\") " pod="openstack/nova-cell0-cell-mapping-52n2r" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.047438 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.048445 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.051063 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.066006 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.110417 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wtct\" (UniqueName: \"kubernetes.io/projected/c3108e0c-07ff-496a-9536-caf85d5dbaba-kube-api-access-8wtct\") pod \"nova-scheduler-0\" (UID: \"c3108e0c-07ff-496a-9536-caf85d5dbaba\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.110475 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-scripts\") pod \"nova-cell0-cell-mapping-52n2r\" (UID: \"66b6902c-c560-43ba-8bf6-b162da3931df\") " pod="openstack/nova-cell0-cell-mapping-52n2r" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.110524 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3108e0c-07ff-496a-9536-caf85d5dbaba-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c3108e0c-07ff-496a-9536-caf85d5dbaba\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.110545 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-config-data\") pod \"nova-cell0-cell-mapping-52n2r\" (UID: \"66b6902c-c560-43ba-8bf6-b162da3931df\") " pod="openstack/nova-cell0-cell-mapping-52n2r" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.110568 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-52n2r\" (UID: \"66b6902c-c560-43ba-8bf6-b162da3931df\") " pod="openstack/nova-cell0-cell-mapping-52n2r" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.110592 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3108e0c-07ff-496a-9536-caf85d5dbaba-config-data\") pod \"nova-scheduler-0\" (UID: \"c3108e0c-07ff-496a-9536-caf85d5dbaba\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.110622 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc8d7\" (UniqueName: \"kubernetes.io/projected/66b6902c-c560-43ba-8bf6-b162da3931df-kube-api-access-zc8d7\") pod \"nova-cell0-cell-mapping-52n2r\" (UID: \"66b6902c-c560-43ba-8bf6-b162da3931df\") " pod="openstack/nova-cell0-cell-mapping-52n2r" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.116993 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-52n2r\" (UID: \"66b6902c-c560-43ba-8bf6-b162da3931df\") " pod="openstack/nova-cell0-cell-mapping-52n2r" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.117554 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-scripts\") pod \"nova-cell0-cell-mapping-52n2r\" (UID: \"66b6902c-c560-43ba-8bf6-b162da3931df\") " pod="openstack/nova-cell0-cell-mapping-52n2r" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.117869 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-config-data\") pod \"nova-cell0-cell-mapping-52n2r\" (UID: \"66b6902c-c560-43ba-8bf6-b162da3931df\") " pod="openstack/nova-cell0-cell-mapping-52n2r" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.133947 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc8d7\" (UniqueName: \"kubernetes.io/projected/66b6902c-c560-43ba-8bf6-b162da3931df-kube-api-access-zc8d7\") pod \"nova-cell0-cell-mapping-52n2r\" (UID: \"66b6902c-c560-43ba-8bf6-b162da3931df\") " pod="openstack/nova-cell0-cell-mapping-52n2r" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.163463 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.164998 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.173874 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.176273 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.201556 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-52n2r" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.212770 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wtct\" (UniqueName: \"kubernetes.io/projected/c3108e0c-07ff-496a-9536-caf85d5dbaba-kube-api-access-8wtct\") pod \"nova-scheduler-0\" (UID: \"c3108e0c-07ff-496a-9536-caf85d5dbaba\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.212909 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3108e0c-07ff-496a-9536-caf85d5dbaba-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c3108e0c-07ff-496a-9536-caf85d5dbaba\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.212969 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3108e0c-07ff-496a-9536-caf85d5dbaba-config-data\") pod \"nova-scheduler-0\" (UID: \"c3108e0c-07ff-496a-9536-caf85d5dbaba\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.216909 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3108e0c-07ff-496a-9536-caf85d5dbaba-config-data\") pod \"nova-scheduler-0\" (UID: \"c3108e0c-07ff-496a-9536-caf85d5dbaba\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.229145 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3108e0c-07ff-496a-9536-caf85d5dbaba-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c3108e0c-07ff-496a-9536-caf85d5dbaba\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.272760 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wtct\" (UniqueName: \"kubernetes.io/projected/c3108e0c-07ff-496a-9536-caf85d5dbaba-kube-api-access-8wtct\") pod \"nova-scheduler-0\" (UID: \"c3108e0c-07ff-496a-9536-caf85d5dbaba\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.302341 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.303917 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.310485 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.315587 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/918450af-cc90-4368-acb7-add3e6d4a737-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"918450af-cc90-4368-acb7-add3e6d4a737\") " pod="openstack/nova-api-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.315812 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/918450af-cc90-4368-acb7-add3e6d4a737-config-data\") pod \"nova-api-0\" (UID: \"918450af-cc90-4368-acb7-add3e6d4a737\") " pod="openstack/nova-api-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.315965 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/918450af-cc90-4368-acb7-add3e6d4a737-logs\") pod \"nova-api-0\" (UID: \"918450af-cc90-4368-acb7-add3e6d4a737\") " pod="openstack/nova-api-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.316097 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbqcc\" (UniqueName: \"kubernetes.io/projected/918450af-cc90-4368-acb7-add3e6d4a737-kube-api-access-wbqcc\") pod \"nova-api-0\" (UID: \"918450af-cc90-4368-acb7-add3e6d4a737\") " pod="openstack/nova-api-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.356326 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.365354 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.381977 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.383574 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.400748 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.419565 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-logs\") pod \"nova-metadata-0\" (UID: \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\") " pod="openstack/nova-metadata-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.419777 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\") " pod="openstack/nova-metadata-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.419849 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/918450af-cc90-4368-acb7-add3e6d4a737-logs\") pod \"nova-api-0\" (UID: \"918450af-cc90-4368-acb7-add3e6d4a737\") " pod="openstack/nova-api-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.419949 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbqcc\" (UniqueName: \"kubernetes.io/projected/918450af-cc90-4368-acb7-add3e6d4a737-kube-api-access-wbqcc\") pod \"nova-api-0\" (UID: \"918450af-cc90-4368-acb7-add3e6d4a737\") " pod="openstack/nova-api-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.420008 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-config-data\") pod \"nova-metadata-0\" (UID: \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\") " pod="openstack/nova-metadata-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.420090 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2zl6\" (UniqueName: \"kubernetes.io/projected/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-kube-api-access-m2zl6\") pod \"nova-metadata-0\" (UID: \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\") " pod="openstack/nova-metadata-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.420126 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/918450af-cc90-4368-acb7-add3e6d4a737-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"918450af-cc90-4368-acb7-add3e6d4a737\") " pod="openstack/nova-api-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.420146 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/918450af-cc90-4368-acb7-add3e6d4a737-config-data\") pod \"nova-api-0\" (UID: \"918450af-cc90-4368-acb7-add3e6d4a737\") " pod="openstack/nova-api-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.428756 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.431374 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/918450af-cc90-4368-acb7-add3e6d4a737-config-data\") pod \"nova-api-0\" (UID: \"918450af-cc90-4368-acb7-add3e6d4a737\") " pod="openstack/nova-api-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.432100 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/918450af-cc90-4368-acb7-add3e6d4a737-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"918450af-cc90-4368-acb7-add3e6d4a737\") " pod="openstack/nova-api-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.443589 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/918450af-cc90-4368-acb7-add3e6d4a737-logs\") pod \"nova-api-0\" (UID: \"918450af-cc90-4368-acb7-add3e6d4a737\") " pod="openstack/nova-api-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.451310 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-f5j6n"] Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.452866 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.454818 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbqcc\" (UniqueName: \"kubernetes.io/projected/918450af-cc90-4368-acb7-add3e6d4a737-kube-api-access-wbqcc\") pod \"nova-api-0\" (UID: \"918450af-cc90-4368-acb7-add3e6d4a737\") " pod="openstack/nova-api-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.465313 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-f5j6n"] Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.521319 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ebe342f-0c82-435c-97f3-3d32d607dae1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2ebe342f-0c82-435c-97f3-3d32d607dae1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.521559 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-logs\") pod \"nova-metadata-0\" (UID: \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\") " pod="openstack/nova-metadata-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.521624 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\") " pod="openstack/nova-metadata-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.521647 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-config\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.521665 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.521688 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.521727 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ebe342f-0c82-435c-97f3-3d32d607dae1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2ebe342f-0c82-435c-97f3-3d32d607dae1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.521759 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.521804 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-config-data\") pod \"nova-metadata-0\" (UID: \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\") " pod="openstack/nova-metadata-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.521824 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7sf8\" (UniqueName: \"kubernetes.io/projected/2ebe342f-0c82-435c-97f3-3d32d607dae1-kube-api-access-v7sf8\") pod \"nova-cell1-novncproxy-0\" (UID: \"2ebe342f-0c82-435c-97f3-3d32d607dae1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.521861 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2zl6\" (UniqueName: \"kubernetes.io/projected/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-kube-api-access-m2zl6\") pod \"nova-metadata-0\" (UID: \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\") " pod="openstack/nova-metadata-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.521892 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcl87\" (UniqueName: \"kubernetes.io/projected/4de0189e-b022-42e1-b572-18e584e0d17c-kube-api-access-zcl87\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.521911 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.522350 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-logs\") pod \"nova-metadata-0\" (UID: \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\") " pod="openstack/nova-metadata-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.526899 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-config-data\") pod \"nova-metadata-0\" (UID: \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\") " pod="openstack/nova-metadata-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.528329 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\") " pod="openstack/nova-metadata-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.544552 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2zl6\" (UniqueName: \"kubernetes.io/projected/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-kube-api-access-m2zl6\") pod \"nova-metadata-0\" (UID: \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\") " pod="openstack/nova-metadata-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.623953 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.624093 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7sf8\" (UniqueName: \"kubernetes.io/projected/2ebe342f-0c82-435c-97f3-3d32d607dae1-kube-api-access-v7sf8\") pod \"nova-cell1-novncproxy-0\" (UID: \"2ebe342f-0c82-435c-97f3-3d32d607dae1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.624215 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcl87\" (UniqueName: \"kubernetes.io/projected/4de0189e-b022-42e1-b572-18e584e0d17c-kube-api-access-zcl87\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.624281 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.624789 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ebe342f-0c82-435c-97f3-3d32d607dae1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2ebe342f-0c82-435c-97f3-3d32d607dae1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.624868 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-config\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.624895 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.624948 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.625034 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ebe342f-0c82-435c-97f3-3d32d607dae1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2ebe342f-0c82-435c-97f3-3d32d607dae1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.625281 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.628399 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-config\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.628779 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.629454 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ebe342f-0c82-435c-97f3-3d32d607dae1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2ebe342f-0c82-435c-97f3-3d32d607dae1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.629599 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.634976 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.635593 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.637967 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ebe342f-0c82-435c-97f3-3d32d607dae1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2ebe342f-0c82-435c-97f3-3d32d607dae1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.642556 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcl87\" (UniqueName: \"kubernetes.io/projected/4de0189e-b022-42e1-b572-18e584e0d17c-kube-api-access-zcl87\") pod \"dnsmasq-dns-845d6d6f59-f5j6n\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.646224 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7sf8\" (UniqueName: \"kubernetes.io/projected/2ebe342f-0c82-435c-97f3-3d32d607dae1-kube-api-access-v7sf8\") pod \"nova-cell1-novncproxy-0\" (UID: \"2ebe342f-0c82-435c-97f3-3d32d607dae1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.654794 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.748656 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.780966 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.868052 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-52n2r"] Feb 27 16:45:34 crc kubenswrapper[4814]: W0227 16:45:34.886110 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66b6902c_c560_43ba_8bf6_b162da3931df.slice/crio-5d8b1acf442f2100ceee895ccf45af67126a7d3289dc13c419b6ca82f6396d83 WatchSource:0}: Error finding container 5d8b1acf442f2100ceee895ccf45af67126a7d3289dc13c419b6ca82f6396d83: Status 404 returned error can't find the container with id 5d8b1acf442f2100ceee895ccf45af67126a7d3289dc13c419b6ca82f6396d83 Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.966621 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-52n2r" event={"ID":"66b6902c-c560-43ba-8bf6-b162da3931df","Type":"ContainerStarted","Data":"5d8b1acf442f2100ceee895ccf45af67126a7d3289dc13c419b6ca82f6396d83"} Feb 27 16:45:34 crc kubenswrapper[4814]: I0227 16:45:34.992778 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 16:45:35 crc kubenswrapper[4814]: W0227 16:45:35.001076 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3108e0c_07ff_496a_9536_caf85d5dbaba.slice/crio-e5f98139b34fe336f002c666d9abbccbea2eb9c6507c52aa051d875a7f86b75c WatchSource:0}: Error finding container e5f98139b34fe336f002c666d9abbccbea2eb9c6507c52aa051d875a7f86b75c: Status 404 returned error can't find the container with id e5f98139b34fe336f002c666d9abbccbea2eb9c6507c52aa051d875a7f86b75c Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.101716 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:45:35 crc kubenswrapper[4814]: W0227 16:45:35.106755 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1e2f7f7_7015_44d4_9938_5a6ce01659ce.slice/crio-085ede0602472f632c055f47e8f35c73096b5eea1032181c68f37e9267352f93 WatchSource:0}: Error finding container 085ede0602472f632c055f47e8f35c73096b5eea1032181c68f37e9267352f93: Status 404 returned error can't find the container with id 085ede0602472f632c055f47e8f35c73096b5eea1032181c68f37e9267352f93 Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.144924 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qm8cq"] Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.147356 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qm8cq" Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.154594 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.154809 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.169507 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qm8cq"] Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.207682 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.258097 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-config-data\") pod \"nova-cell1-conductor-db-sync-qm8cq\" (UID: \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\") " pod="openstack/nova-cell1-conductor-db-sync-qm8cq" Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.258643 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st6rb\" (UniqueName: \"kubernetes.io/projected/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-kube-api-access-st6rb\") pod \"nova-cell1-conductor-db-sync-qm8cq\" (UID: \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\") " pod="openstack/nova-cell1-conductor-db-sync-qm8cq" Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.258774 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-qm8cq\" (UID: \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\") " pod="openstack/nova-cell1-conductor-db-sync-qm8cq" Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.259019 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-scripts\") pod \"nova-cell1-conductor-db-sync-qm8cq\" (UID: \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\") " pod="openstack/nova-cell1-conductor-db-sync-qm8cq" Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.361150 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-config-data\") pod \"nova-cell1-conductor-db-sync-qm8cq\" (UID: \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\") " pod="openstack/nova-cell1-conductor-db-sync-qm8cq" Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.361242 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st6rb\" (UniqueName: \"kubernetes.io/projected/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-kube-api-access-st6rb\") pod \"nova-cell1-conductor-db-sync-qm8cq\" (UID: \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\") " pod="openstack/nova-cell1-conductor-db-sync-qm8cq" Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.361342 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-qm8cq\" (UID: \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\") " pod="openstack/nova-cell1-conductor-db-sync-qm8cq" Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.361407 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-scripts\") pod \"nova-cell1-conductor-db-sync-qm8cq\" (UID: \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\") " pod="openstack/nova-cell1-conductor-db-sync-qm8cq" Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.367033 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-qm8cq\" (UID: \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\") " pod="openstack/nova-cell1-conductor-db-sync-qm8cq" Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.368032 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-scripts\") pod \"nova-cell1-conductor-db-sync-qm8cq\" (UID: \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\") " pod="openstack/nova-cell1-conductor-db-sync-qm8cq" Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.375232 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-config-data\") pod \"nova-cell1-conductor-db-sync-qm8cq\" (UID: \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\") " pod="openstack/nova-cell1-conductor-db-sync-qm8cq" Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.380155 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st6rb\" (UniqueName: \"kubernetes.io/projected/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-kube-api-access-st6rb\") pod \"nova-cell1-conductor-db-sync-qm8cq\" (UID: \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\") " pod="openstack/nova-cell1-conductor-db-sync-qm8cq" Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.412480 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.438395 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-f5j6n"] Feb 27 16:45:35 crc kubenswrapper[4814]: W0227 16:45:35.444634 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4de0189e_b022_42e1_b572_18e584e0d17c.slice/crio-3bca4e9419e6621c53311e778eb9d949ceb4eb6bcc0023469f13dd7fbb8e18a3 WatchSource:0}: Error finding container 3bca4e9419e6621c53311e778eb9d949ceb4eb6bcc0023469f13dd7fbb8e18a3: Status 404 returned error can't find the container with id 3bca4e9419e6621c53311e778eb9d949ceb4eb6bcc0023469f13dd7fbb8e18a3 Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.555826 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qm8cq" Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.981421 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c3108e0c-07ff-496a-9536-caf85d5dbaba","Type":"ContainerStarted","Data":"e5f98139b34fe336f002c666d9abbccbea2eb9c6507c52aa051d875a7f86b75c"} Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.989009 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"918450af-cc90-4368-acb7-add3e6d4a737","Type":"ContainerStarted","Data":"ed46c72ce348ba9ba7c3f2fe840740b26fa676e8403ed5053a3ff15ae8536005"} Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.991243 4814 generic.go:334] "Generic (PLEG): container finished" podID="4de0189e-b022-42e1-b572-18e584e0d17c" containerID="302cb5ab87f1e8d112c3e97fd57c98544f5a22484607f06843605a6d04860365" exitCode=0 Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.991307 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" event={"ID":"4de0189e-b022-42e1-b572-18e584e0d17c","Type":"ContainerDied","Data":"302cb5ab87f1e8d112c3e97fd57c98544f5a22484607f06843605a6d04860365"} Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.991323 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" event={"ID":"4de0189e-b022-42e1-b572-18e584e0d17c","Type":"ContainerStarted","Data":"3bca4e9419e6621c53311e778eb9d949ceb4eb6bcc0023469f13dd7fbb8e18a3"} Feb 27 16:45:35 crc kubenswrapper[4814]: I0227 16:45:35.995947 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2ebe342f-0c82-435c-97f3-3d32d607dae1","Type":"ContainerStarted","Data":"abba409ed86941f8b4ba3cd6d005ebed1563192b18590bbb6dd77cf8a2568467"} Feb 27 16:45:36 crc kubenswrapper[4814]: I0227 16:45:36.010682 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-52n2r" event={"ID":"66b6902c-c560-43ba-8bf6-b162da3931df","Type":"ContainerStarted","Data":"838f5dea6dde2748cbb8c5f618de14790336077931da329733c57a834e12c718"} Feb 27 16:45:36 crc kubenswrapper[4814]: I0227 16:45:36.012678 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b1e2f7f7-7015-44d4-9938-5a6ce01659ce","Type":"ContainerStarted","Data":"085ede0602472f632c055f47e8f35c73096b5eea1032181c68f37e9267352f93"} Feb 27 16:45:36 crc kubenswrapper[4814]: I0227 16:45:36.050588 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-52n2r" podStartSLOduration=3.050567996 podStartE2EDuration="3.050567996s" podCreationTimestamp="2026-02-27 16:45:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:45:36.041315996 +0000 UTC m=+1348.493940866" watchObservedRunningTime="2026-02-27 16:45:36.050567996 +0000 UTC m=+1348.503192826" Feb 27 16:45:36 crc kubenswrapper[4814]: I0227 16:45:36.093305 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qm8cq"] Feb 27 16:45:36 crc kubenswrapper[4814]: W0227 16:45:36.098490 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8be859d_3be2_4c3f_b10e_b0d04d8fe14d.slice/crio-19476a39fed26e2f40be51c86e7d4895cf817cc64dee97e8b19e89c282a644d7 WatchSource:0}: Error finding container 19476a39fed26e2f40be51c86e7d4895cf817cc64dee97e8b19e89c282a644d7: Status 404 returned error can't find the container with id 19476a39fed26e2f40be51c86e7d4895cf817cc64dee97e8b19e89c282a644d7 Feb 27 16:45:37 crc kubenswrapper[4814]: I0227 16:45:37.025387 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qm8cq" event={"ID":"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d","Type":"ContainerStarted","Data":"ca4d6256ededf270e247f60cc5ccd983bac59c8fc25d2020e5e5bbdc188cf590"} Feb 27 16:45:37 crc kubenswrapper[4814]: I0227 16:45:37.025833 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qm8cq" event={"ID":"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d","Type":"ContainerStarted","Data":"19476a39fed26e2f40be51c86e7d4895cf817cc64dee97e8b19e89c282a644d7"} Feb 27 16:45:37 crc kubenswrapper[4814]: I0227 16:45:37.028982 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" event={"ID":"4de0189e-b022-42e1-b572-18e584e0d17c","Type":"ContainerStarted","Data":"0a6fb20559f0d4da93935d261d6ab12128c9cb409086af7e25a93969e2cee4f4"} Feb 27 16:45:37 crc kubenswrapper[4814]: I0227 16:45:37.029133 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:37 crc kubenswrapper[4814]: I0227 16:45:37.083611 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" podStartSLOduration=3.083590556 podStartE2EDuration="3.083590556s" podCreationTimestamp="2026-02-27 16:45:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:45:37.080100047 +0000 UTC m=+1349.532724877" watchObservedRunningTime="2026-02-27 16:45:37.083590556 +0000 UTC m=+1349.536215386" Feb 27 16:45:37 crc kubenswrapper[4814]: I0227 16:45:37.089874 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-qm8cq" podStartSLOduration=2.089853693 podStartE2EDuration="2.089853693s" podCreationTimestamp="2026-02-27 16:45:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:45:37.058236603 +0000 UTC m=+1349.510861433" watchObservedRunningTime="2026-02-27 16:45:37.089853693 +0000 UTC m=+1349.542478523" Feb 27 16:45:37 crc kubenswrapper[4814]: I0227 16:45:37.543142 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:45:37 crc kubenswrapper[4814]: I0227 16:45:37.550691 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.052186 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b1e2f7f7-7015-44d4-9938-5a6ce01659ce","Type":"ContainerStarted","Data":"80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04"} Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.052642 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b1e2f7f7-7015-44d4-9938-5a6ce01659ce","Type":"ContainerStarted","Data":"d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34"} Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.052890 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b1e2f7f7-7015-44d4-9938-5a6ce01659ce" containerName="nova-metadata-log" containerID="cri-o://d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34" gracePeriod=30 Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.053894 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b1e2f7f7-7015-44d4-9938-5a6ce01659ce" containerName="nova-metadata-metadata" containerID="cri-o://80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04" gracePeriod=30 Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.064660 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c3108e0c-07ff-496a-9536-caf85d5dbaba","Type":"ContainerStarted","Data":"17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a"} Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.068219 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"918450af-cc90-4368-acb7-add3e6d4a737","Type":"ContainerStarted","Data":"97899ac9abe6188d7bb13926164f0fe274525899a6ff5367f2c88ff54c05ea41"} Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.068246 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"918450af-cc90-4368-acb7-add3e6d4a737","Type":"ContainerStarted","Data":"17c52fa0571b3a706dba4f2fa1daa148aa265d65a8e3de3357e10d43af9cc9ea"} Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.070080 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2ebe342f-0c82-435c-97f3-3d32d607dae1","Type":"ContainerStarted","Data":"b7f23a798416a3ef7914d9218aab56b18b95ad2543be35aa29983c847a419d3d"} Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.070174 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="2ebe342f-0c82-435c-97f3-3d32d607dae1" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://b7f23a798416a3ef7914d9218aab56b18b95ad2543be35aa29983c847a419d3d" gracePeriod=30 Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.088420 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.236833205 podStartE2EDuration="5.08839135s" podCreationTimestamp="2026-02-27 16:45:34 +0000 UTC" firstStartedPulling="2026-02-27 16:45:35.110056291 +0000 UTC m=+1347.562681121" lastFinishedPulling="2026-02-27 16:45:37.961614436 +0000 UTC m=+1350.414239266" observedRunningTime="2026-02-27 16:45:39.082680372 +0000 UTC m=+1351.535305242" watchObservedRunningTime="2026-02-27 16:45:39.08839135 +0000 UTC m=+1351.541016200" Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.124771 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.167998222 podStartE2EDuration="5.124723527s" podCreationTimestamp="2026-02-27 16:45:34 +0000 UTC" firstStartedPulling="2026-02-27 16:45:35.004886891 +0000 UTC m=+1347.457511731" lastFinishedPulling="2026-02-27 16:45:37.961612166 +0000 UTC m=+1350.414237036" observedRunningTime="2026-02-27 16:45:39.109642435 +0000 UTC m=+1351.562267285" watchObservedRunningTime="2026-02-27 16:45:39.124723527 +0000 UTC m=+1351.577348357" Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.133628 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.591231971 podStartE2EDuration="5.133586014s" podCreationTimestamp="2026-02-27 16:45:34 +0000 UTC" firstStartedPulling="2026-02-27 16:45:35.420151142 +0000 UTC m=+1347.872775972" lastFinishedPulling="2026-02-27 16:45:37.962505195 +0000 UTC m=+1350.415130015" observedRunningTime="2026-02-27 16:45:39.125868963 +0000 UTC m=+1351.578493803" watchObservedRunningTime="2026-02-27 16:45:39.133586014 +0000 UTC m=+1351.586210844" Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.146531 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.371189639 podStartE2EDuration="5.146512249s" podCreationTimestamp="2026-02-27 16:45:34 +0000 UTC" firstStartedPulling="2026-02-27 16:45:35.208903824 +0000 UTC m=+1347.661528654" lastFinishedPulling="2026-02-27 16:45:37.984226414 +0000 UTC m=+1350.436851264" observedRunningTime="2026-02-27 16:45:39.141579714 +0000 UTC m=+1351.594204564" watchObservedRunningTime="2026-02-27 16:45:39.146512249 +0000 UTC m=+1351.599137089" Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.366956 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 27 16:45:39 crc kubenswrapper[4814]: E0227 16:45:39.613594 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa51609e_188a_4ded_962b_9f9641404afc.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa51609e_188a_4ded_962b_9f9641404afc.slice/crio-eb7a32b82b3fba9fb45c72f26ffdb9697916b2788f479c2eb70fe921070b3abe\": RecentStats: unable to find data in memory cache]" Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.655698 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.655758 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.749848 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.753638 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.880124 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2zl6\" (UniqueName: \"kubernetes.io/projected/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-kube-api-access-m2zl6\") pod \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\" (UID: \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\") " Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.881612 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-combined-ca-bundle\") pod \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\" (UID: \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\") " Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.881790 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-logs\") pod \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\" (UID: \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\") " Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.881837 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-config-data\") pod \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\" (UID: \"b1e2f7f7-7015-44d4-9938-5a6ce01659ce\") " Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.882043 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-logs" (OuterVolumeSpecName: "logs") pod "b1e2f7f7-7015-44d4-9938-5a6ce01659ce" (UID: "b1e2f7f7-7015-44d4-9938-5a6ce01659ce"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.882636 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.888273 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-kube-api-access-m2zl6" (OuterVolumeSpecName: "kube-api-access-m2zl6") pod "b1e2f7f7-7015-44d4-9938-5a6ce01659ce" (UID: "b1e2f7f7-7015-44d4-9938-5a6ce01659ce"). InnerVolumeSpecName "kube-api-access-m2zl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.910369 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-config-data" (OuterVolumeSpecName: "config-data") pod "b1e2f7f7-7015-44d4-9938-5a6ce01659ce" (UID: "b1e2f7f7-7015-44d4-9938-5a6ce01659ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.916412 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1e2f7f7-7015-44d4-9938-5a6ce01659ce" (UID: "b1e2f7f7-7015-44d4-9938-5a6ce01659ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.984848 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2zl6\" (UniqueName: \"kubernetes.io/projected/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-kube-api-access-m2zl6\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.984897 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:39 crc kubenswrapper[4814]: I0227 16:45:39.984910 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1e2f7f7-7015-44d4-9938-5a6ce01659ce-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.082326 4814 generic.go:334] "Generic (PLEG): container finished" podID="b1e2f7f7-7015-44d4-9938-5a6ce01659ce" containerID="80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04" exitCode=0 Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.082359 4814 generic.go:334] "Generic (PLEG): container finished" podID="b1e2f7f7-7015-44d4-9938-5a6ce01659ce" containerID="d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34" exitCode=143 Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.082428 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.082492 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b1e2f7f7-7015-44d4-9938-5a6ce01659ce","Type":"ContainerDied","Data":"80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04"} Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.082537 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b1e2f7f7-7015-44d4-9938-5a6ce01659ce","Type":"ContainerDied","Data":"d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34"} Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.082557 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b1e2f7f7-7015-44d4-9938-5a6ce01659ce","Type":"ContainerDied","Data":"085ede0602472f632c055f47e8f35c73096b5eea1032181c68f37e9267352f93"} Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.082582 4814 scope.go:117] "RemoveContainer" containerID="80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.118583 4814 scope.go:117] "RemoveContainer" containerID="d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.133732 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.146959 4814 scope.go:117] "RemoveContainer" containerID="80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04" Feb 27 16:45:40 crc kubenswrapper[4814]: E0227 16:45:40.147451 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04\": container with ID starting with 80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04 not found: ID does not exist" containerID="80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.147489 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04"} err="failed to get container status \"80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04\": rpc error: code = NotFound desc = could not find container \"80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04\": container with ID starting with 80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04 not found: ID does not exist" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.147514 4814 scope.go:117] "RemoveContainer" containerID="d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34" Feb 27 16:45:40 crc kubenswrapper[4814]: E0227 16:45:40.147878 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34\": container with ID starting with d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34 not found: ID does not exist" containerID="d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.147900 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34"} err="failed to get container status \"d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34\": rpc error: code = NotFound desc = could not find container \"d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34\": container with ID starting with d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34 not found: ID does not exist" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.147916 4814 scope.go:117] "RemoveContainer" containerID="80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.149896 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04"} err="failed to get container status \"80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04\": rpc error: code = NotFound desc = could not find container \"80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04\": container with ID starting with 80ec7ff189dd0b036a09906199e4d0663e47cdc872fa5e69bab30509b1f0ed04 not found: ID does not exist" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.149939 4814 scope.go:117] "RemoveContainer" containerID="d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.150377 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34"} err="failed to get container status \"d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34\": rpc error: code = NotFound desc = could not find container \"d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34\": container with ID starting with d720a1f47f0f761a20d1dcbaafabe28acb221ae7acb7d38e4262c5949d233c34 not found: ID does not exist" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.151571 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.162441 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:45:40 crc kubenswrapper[4814]: E0227 16:45:40.162820 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e2f7f7-7015-44d4-9938-5a6ce01659ce" containerName="nova-metadata-metadata" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.162837 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e2f7f7-7015-44d4-9938-5a6ce01659ce" containerName="nova-metadata-metadata" Feb 27 16:45:40 crc kubenswrapper[4814]: E0227 16:45:40.162874 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e2f7f7-7015-44d4-9938-5a6ce01659ce" containerName="nova-metadata-log" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.162881 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e2f7f7-7015-44d4-9938-5a6ce01659ce" containerName="nova-metadata-log" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.163060 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1e2f7f7-7015-44d4-9938-5a6ce01659ce" containerName="nova-metadata-log" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.163075 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1e2f7f7-7015-44d4-9938-5a6ce01659ce" containerName="nova-metadata-metadata" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.164006 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.167176 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.173176 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.189231 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.293961 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5rbq\" (UniqueName: \"kubernetes.io/projected/27249d07-0ac4-49c6-9040-2fb2f22a59fd-kube-api-access-d5rbq\") pod \"nova-metadata-0\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.294159 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.294204 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.294285 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-config-data\") pod \"nova-metadata-0\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.294590 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27249d07-0ac4-49c6-9040-2fb2f22a59fd-logs\") pod \"nova-metadata-0\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.397329 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27249d07-0ac4-49c6-9040-2fb2f22a59fd-logs\") pod \"nova-metadata-0\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.397454 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5rbq\" (UniqueName: \"kubernetes.io/projected/27249d07-0ac4-49c6-9040-2fb2f22a59fd-kube-api-access-d5rbq\") pod \"nova-metadata-0\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.397582 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.397629 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.397730 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-config-data\") pod \"nova-metadata-0\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.398058 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27249d07-0ac4-49c6-9040-2fb2f22a59fd-logs\") pod \"nova-metadata-0\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.403375 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-config-data\") pod \"nova-metadata-0\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.409185 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.410080 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.431860 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5rbq\" (UniqueName: \"kubernetes.io/projected/27249d07-0ac4-49c6-9040-2fb2f22a59fd-kube-api-access-d5rbq\") pod \"nova-metadata-0\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.483774 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 16:45:40 crc kubenswrapper[4814]: I0227 16:45:40.524008 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1e2f7f7-7015-44d4-9938-5a6ce01659ce" path="/var/lib/kubelet/pods/b1e2f7f7-7015-44d4-9938-5a6ce01659ce/volumes" Feb 27 16:45:41 crc kubenswrapper[4814]: I0227 16:45:41.034643 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:45:41 crc kubenswrapper[4814]: I0227 16:45:41.094226 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"27249d07-0ac4-49c6-9040-2fb2f22a59fd","Type":"ContainerStarted","Data":"28fdfdfa2beea9a3fc22f2ac21699ccc3989ab9525253999bf6601642f5dfa8f"} Feb 27 16:45:42 crc kubenswrapper[4814]: I0227 16:45:42.116749 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"27249d07-0ac4-49c6-9040-2fb2f22a59fd","Type":"ContainerStarted","Data":"730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830"} Feb 27 16:45:42 crc kubenswrapper[4814]: I0227 16:45:42.117086 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"27249d07-0ac4-49c6-9040-2fb2f22a59fd","Type":"ContainerStarted","Data":"2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5"} Feb 27 16:45:42 crc kubenswrapper[4814]: I0227 16:45:42.148920 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.148899425 podStartE2EDuration="2.148899425s" podCreationTimestamp="2026-02-27 16:45:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:45:42.138356315 +0000 UTC m=+1354.590981185" watchObservedRunningTime="2026-02-27 16:45:42.148899425 +0000 UTC m=+1354.601524255" Feb 27 16:45:43 crc kubenswrapper[4814]: I0227 16:45:43.127979 4814 generic.go:334] "Generic (PLEG): container finished" podID="66b6902c-c560-43ba-8bf6-b162da3931df" containerID="838f5dea6dde2748cbb8c5f618de14790336077931da329733c57a834e12c718" exitCode=0 Feb 27 16:45:43 crc kubenswrapper[4814]: I0227 16:45:43.128076 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-52n2r" event={"ID":"66b6902c-c560-43ba-8bf6-b162da3931df","Type":"ContainerDied","Data":"838f5dea6dde2748cbb8c5f618de14790336077931da329733c57a834e12c718"} Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.366577 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.423870 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.629911 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.629977 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.642541 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-52n2r" Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.701156 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-scripts\") pod \"66b6902c-c560-43ba-8bf6-b162da3931df\" (UID: \"66b6902c-c560-43ba-8bf6-b162da3931df\") " Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.701242 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zc8d7\" (UniqueName: \"kubernetes.io/projected/66b6902c-c560-43ba-8bf6-b162da3931df-kube-api-access-zc8d7\") pod \"66b6902c-c560-43ba-8bf6-b162da3931df\" (UID: \"66b6902c-c560-43ba-8bf6-b162da3931df\") " Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.701295 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-config-data\") pod \"66b6902c-c560-43ba-8bf6-b162da3931df\" (UID: \"66b6902c-c560-43ba-8bf6-b162da3931df\") " Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.701365 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-combined-ca-bundle\") pod \"66b6902c-c560-43ba-8bf6-b162da3931df\" (UID: \"66b6902c-c560-43ba-8bf6-b162da3931df\") " Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.712024 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66b6902c-c560-43ba-8bf6-b162da3931df-kube-api-access-zc8d7" (OuterVolumeSpecName: "kube-api-access-zc8d7") pod "66b6902c-c560-43ba-8bf6-b162da3931df" (UID: "66b6902c-c560-43ba-8bf6-b162da3931df"). InnerVolumeSpecName "kube-api-access-zc8d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.724217 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-scripts" (OuterVolumeSpecName: "scripts") pod "66b6902c-c560-43ba-8bf6-b162da3931df" (UID: "66b6902c-c560-43ba-8bf6-b162da3931df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.753458 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66b6902c-c560-43ba-8bf6-b162da3931df" (UID: "66b6902c-c560-43ba-8bf6-b162da3931df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.770749 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-config-data" (OuterVolumeSpecName: "config-data") pod "66b6902c-c560-43ba-8bf6-b162da3931df" (UID: "66b6902c-c560-43ba-8bf6-b162da3931df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.783613 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.821181 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.821220 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.821234 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66b6902c-c560-43ba-8bf6-b162da3931df-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.821243 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zc8d7\" (UniqueName: \"kubernetes.io/projected/66b6902c-c560-43ba-8bf6-b162da3931df-kube-api-access-zc8d7\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.875815 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-frg9p"] Feb 27 16:45:44 crc kubenswrapper[4814]: I0227 16:45:44.876088 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-frg9p" podUID="0f95f00d-d69f-4734-ba33-436879200784" containerName="dnsmasq-dns" containerID="cri-o://5dfc7a6c8a51195c471ca7fcf381df65e5441358bcbf95788bec2b01f67f6444" gracePeriod=10 Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.150761 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-52n2r" event={"ID":"66b6902c-c560-43ba-8bf6-b162da3931df","Type":"ContainerDied","Data":"5d8b1acf442f2100ceee895ccf45af67126a7d3289dc13c419b6ca82f6396d83"} Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.150797 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d8b1acf442f2100ceee895ccf45af67126a7d3289dc13c419b6ca82f6396d83" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.150855 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-52n2r" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.164814 4814 generic.go:334] "Generic (PLEG): container finished" podID="0f95f00d-d69f-4734-ba33-436879200784" containerID="5dfc7a6c8a51195c471ca7fcf381df65e5441358bcbf95788bec2b01f67f6444" exitCode=0 Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.164889 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-frg9p" event={"ID":"0f95f00d-d69f-4734-ba33-436879200784","Type":"ContainerDied","Data":"5dfc7a6c8a51195c471ca7fcf381df65e5441358bcbf95788bec2b01f67f6444"} Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.169896 4814 generic.go:334] "Generic (PLEG): container finished" podID="f8be859d-3be2-4c3f-b10e-b0d04d8fe14d" containerID="ca4d6256ededf270e247f60cc5ccd983bac59c8fc25d2020e5e5bbdc188cf590" exitCode=0 Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.170869 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qm8cq" event={"ID":"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d","Type":"ContainerDied","Data":"ca4d6256ededf270e247f60cc5ccd983bac59c8fc25d2020e5e5bbdc188cf590"} Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.223354 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.317605 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.419614 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.419817 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="918450af-cc90-4368-acb7-add3e6d4a737" containerName="nova-api-log" containerID="cri-o://17c52fa0571b3a706dba4f2fa1daa148aa265d65a8e3de3357e10d43af9cc9ea" gracePeriod=30 Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.419916 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="918450af-cc90-4368-acb7-add3e6d4a737" containerName="nova-api-api" containerID="cri-o://97899ac9abe6188d7bb13926164f0fe274525899a6ff5367f2c88ff54c05ea41" gracePeriod=30 Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.424721 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="918450af-cc90-4368-acb7-add3e6d4a737" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": EOF" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.424806 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="918450af-cc90-4368-acb7-add3e6d4a737" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": EOF" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.431855 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-ovsdbserver-sb\") pod \"0f95f00d-d69f-4734-ba33-436879200784\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.431904 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj995\" (UniqueName: \"kubernetes.io/projected/0f95f00d-d69f-4734-ba33-436879200784-kube-api-access-nj995\") pod \"0f95f00d-d69f-4734-ba33-436879200784\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.432026 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-dns-swift-storage-0\") pod \"0f95f00d-d69f-4734-ba33-436879200784\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.432083 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-config\") pod \"0f95f00d-d69f-4734-ba33-436879200784\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.432140 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-ovsdbserver-nb\") pod \"0f95f00d-d69f-4734-ba33-436879200784\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.432192 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-dns-svc\") pod \"0f95f00d-d69f-4734-ba33-436879200784\" (UID: \"0f95f00d-d69f-4734-ba33-436879200784\") " Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.458858 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f95f00d-d69f-4734-ba33-436879200784-kube-api-access-nj995" (OuterVolumeSpecName: "kube-api-access-nj995") pod "0f95f00d-d69f-4734-ba33-436879200784" (UID: "0f95f00d-d69f-4734-ba33-436879200784"). InnerVolumeSpecName "kube-api-access-nj995". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.477291 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.477512 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="27249d07-0ac4-49c6-9040-2fb2f22a59fd" containerName="nova-metadata-log" containerID="cri-o://2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5" gracePeriod=30 Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.477655 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="27249d07-0ac4-49c6-9040-2fb2f22a59fd" containerName="nova-metadata-metadata" containerID="cri-o://730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830" gracePeriod=30 Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.484988 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.485026 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.494902 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0f95f00d-d69f-4734-ba33-436879200784" (UID: "0f95f00d-d69f-4734-ba33-436879200784"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.511716 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0f95f00d-d69f-4734-ba33-436879200784" (UID: "0f95f00d-d69f-4734-ba33-436879200784"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.520240 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0f95f00d-d69f-4734-ba33-436879200784" (UID: "0f95f00d-d69f-4734-ba33-436879200784"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.527088 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0f95f00d-d69f-4734-ba33-436879200784" (UID: "0f95f00d-d69f-4734-ba33-436879200784"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.532903 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-config" (OuterVolumeSpecName: "config") pod "0f95f00d-d69f-4734-ba33-436879200784" (UID: "0f95f00d-d69f-4734-ba33-436879200784"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.534080 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.534102 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj995\" (UniqueName: \"kubernetes.io/projected/0f95f00d-d69f-4734-ba33-436879200784-kube-api-access-nj995\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.534114 4814 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.534124 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.534133 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.534142 4814 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f95f00d-d69f-4734-ba33-436879200784-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:45 crc kubenswrapper[4814]: I0227 16:45:45.675759 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.087989 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.144732 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5rbq\" (UniqueName: \"kubernetes.io/projected/27249d07-0ac4-49c6-9040-2fb2f22a59fd-kube-api-access-d5rbq\") pod \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.145054 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-nova-metadata-tls-certs\") pod \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.145091 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27249d07-0ac4-49c6-9040-2fb2f22a59fd-logs\") pod \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.145139 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-config-data\") pod \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.145211 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-combined-ca-bundle\") pod \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\" (UID: \"27249d07-0ac4-49c6-9040-2fb2f22a59fd\") " Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.145844 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27249d07-0ac4-49c6-9040-2fb2f22a59fd-logs" (OuterVolumeSpecName: "logs") pod "27249d07-0ac4-49c6-9040-2fb2f22a59fd" (UID: "27249d07-0ac4-49c6-9040-2fb2f22a59fd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.150873 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27249d07-0ac4-49c6-9040-2fb2f22a59fd-kube-api-access-d5rbq" (OuterVolumeSpecName: "kube-api-access-d5rbq") pod "27249d07-0ac4-49c6-9040-2fb2f22a59fd" (UID: "27249d07-0ac4-49c6-9040-2fb2f22a59fd"). InnerVolumeSpecName "kube-api-access-d5rbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.171611 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-config-data" (OuterVolumeSpecName: "config-data") pod "27249d07-0ac4-49c6-9040-2fb2f22a59fd" (UID: "27249d07-0ac4-49c6-9040-2fb2f22a59fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.180446 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27249d07-0ac4-49c6-9040-2fb2f22a59fd" (UID: "27249d07-0ac4-49c6-9040-2fb2f22a59fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.181340 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-frg9p" event={"ID":"0f95f00d-d69f-4734-ba33-436879200784","Type":"ContainerDied","Data":"72ab4abb88e1f8871ee0a6289805887ab2cac8acd6c0b2a2e1f46bb4da914ae6"} Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.181372 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-frg9p" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.181395 4814 scope.go:117] "RemoveContainer" containerID="5dfc7a6c8a51195c471ca7fcf381df65e5441358bcbf95788bec2b01f67f6444" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.184949 4814 generic.go:334] "Generic (PLEG): container finished" podID="918450af-cc90-4368-acb7-add3e6d4a737" containerID="17c52fa0571b3a706dba4f2fa1daa148aa265d65a8e3de3357e10d43af9cc9ea" exitCode=143 Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.185027 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"918450af-cc90-4368-acb7-add3e6d4a737","Type":"ContainerDied","Data":"17c52fa0571b3a706dba4f2fa1daa148aa265d65a8e3de3357e10d43af9cc9ea"} Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.187905 4814 generic.go:334] "Generic (PLEG): container finished" podID="27249d07-0ac4-49c6-9040-2fb2f22a59fd" containerID="730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830" exitCode=0 Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.187933 4814 generic.go:334] "Generic (PLEG): container finished" podID="27249d07-0ac4-49c6-9040-2fb2f22a59fd" containerID="2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5" exitCode=143 Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.187992 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"27249d07-0ac4-49c6-9040-2fb2f22a59fd","Type":"ContainerDied","Data":"730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830"} Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.188054 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"27249d07-0ac4-49c6-9040-2fb2f22a59fd","Type":"ContainerDied","Data":"2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5"} Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.188072 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"27249d07-0ac4-49c6-9040-2fb2f22a59fd","Type":"ContainerDied","Data":"28fdfdfa2beea9a3fc22f2ac21699ccc3989ab9525253999bf6601642f5dfa8f"} Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.188214 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.218995 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "27249d07-0ac4-49c6-9040-2fb2f22a59fd" (UID: "27249d07-0ac4-49c6-9040-2fb2f22a59fd"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.232329 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-frg9p"] Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.238686 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-frg9p"] Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.246096 4814 scope.go:117] "RemoveContainer" containerID="7180d59d8e49f1f7359649d5407887c77fe44265b021497b2452b85f7866d457" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.247809 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5rbq\" (UniqueName: \"kubernetes.io/projected/27249d07-0ac4-49c6-9040-2fb2f22a59fd-kube-api-access-d5rbq\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.247843 4814 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.248030 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27249d07-0ac4-49c6-9040-2fb2f22a59fd-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.248045 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.248058 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27249d07-0ac4-49c6-9040-2fb2f22a59fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.272079 4814 scope.go:117] "RemoveContainer" containerID="730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.310078 4814 scope.go:117] "RemoveContainer" containerID="2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.354115 4814 scope.go:117] "RemoveContainer" containerID="730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830" Feb 27 16:45:46 crc kubenswrapper[4814]: E0227 16:45:46.354603 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830\": container with ID starting with 730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830 not found: ID does not exist" containerID="730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.354646 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830"} err="failed to get container status \"730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830\": rpc error: code = NotFound desc = could not find container \"730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830\": container with ID starting with 730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830 not found: ID does not exist" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.354674 4814 scope.go:117] "RemoveContainer" containerID="2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5" Feb 27 16:45:46 crc kubenswrapper[4814]: E0227 16:45:46.355035 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5\": container with ID starting with 2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5 not found: ID does not exist" containerID="2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.355060 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5"} err="failed to get container status \"2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5\": rpc error: code = NotFound desc = could not find container \"2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5\": container with ID starting with 2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5 not found: ID does not exist" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.355077 4814 scope.go:117] "RemoveContainer" containerID="730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.367473 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830"} err="failed to get container status \"730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830\": rpc error: code = NotFound desc = could not find container \"730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830\": container with ID starting with 730ba841974c482aca56691594d48eb04a7e37619976750cf640f9f46c227830 not found: ID does not exist" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.367523 4814 scope.go:117] "RemoveContainer" containerID="2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.370329 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5"} err="failed to get container status \"2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5\": rpc error: code = NotFound desc = could not find container \"2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5\": container with ID starting with 2dca2590eda9365530e35638dd9d740ecde4309cbc15f6676bed0df0627395b5 not found: ID does not exist" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.458046 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qm8cq" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.501521 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f95f00d-d69f-4734-ba33-436879200784" path="/var/lib/kubelet/pods/0f95f00d-d69f-4734-ba33-436879200784/volumes" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.532464 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.546335 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.553122 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st6rb\" (UniqueName: \"kubernetes.io/projected/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-kube-api-access-st6rb\") pod \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\" (UID: \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\") " Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.553276 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-config-data\") pod \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\" (UID: \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\") " Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.553468 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-scripts\") pod \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\" (UID: \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\") " Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.553508 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-combined-ca-bundle\") pod \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\" (UID: \"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d\") " Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.557852 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:45:46 crc kubenswrapper[4814]: E0227 16:45:46.558573 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27249d07-0ac4-49c6-9040-2fb2f22a59fd" containerName="nova-metadata-log" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.559227 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="27249d07-0ac4-49c6-9040-2fb2f22a59fd" containerName="nova-metadata-log" Feb 27 16:45:46 crc kubenswrapper[4814]: E0227 16:45:46.560193 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f95f00d-d69f-4734-ba33-436879200784" containerName="dnsmasq-dns" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.560351 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f95f00d-d69f-4734-ba33-436879200784" containerName="dnsmasq-dns" Feb 27 16:45:46 crc kubenswrapper[4814]: E0227 16:45:46.560455 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27249d07-0ac4-49c6-9040-2fb2f22a59fd" containerName="nova-metadata-metadata" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.560529 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="27249d07-0ac4-49c6-9040-2fb2f22a59fd" containerName="nova-metadata-metadata" Feb 27 16:45:46 crc kubenswrapper[4814]: E0227 16:45:46.560616 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b6902c-c560-43ba-8bf6-b162da3931df" containerName="nova-manage" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.560691 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b6902c-c560-43ba-8bf6-b162da3931df" containerName="nova-manage" Feb 27 16:45:46 crc kubenswrapper[4814]: E0227 16:45:46.560782 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f95f00d-d69f-4734-ba33-436879200784" containerName="init" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.560859 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f95f00d-d69f-4734-ba33-436879200784" containerName="init" Feb 27 16:45:46 crc kubenswrapper[4814]: E0227 16:45:46.561026 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8be859d-3be2-4c3f-b10e-b0d04d8fe14d" containerName="nova-cell1-conductor-db-sync" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.561107 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8be859d-3be2-4c3f-b10e-b0d04d8fe14d" containerName="nova-cell1-conductor-db-sync" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.561422 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="27249d07-0ac4-49c6-9040-2fb2f22a59fd" containerName="nova-metadata-metadata" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.561518 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f95f00d-d69f-4734-ba33-436879200784" containerName="dnsmasq-dns" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.561594 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8be859d-3be2-4c3f-b10e-b0d04d8fe14d" containerName="nova-cell1-conductor-db-sync" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.561705 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b6902c-c560-43ba-8bf6-b162da3931df" containerName="nova-manage" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.561792 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="27249d07-0ac4-49c6-9040-2fb2f22a59fd" containerName="nova-metadata-log" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.563634 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.564908 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.558055 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-kube-api-access-st6rb" (OuterVolumeSpecName: "kube-api-access-st6rb") pod "f8be859d-3be2-4c3f-b10e-b0d04d8fe14d" (UID: "f8be859d-3be2-4c3f-b10e-b0d04d8fe14d"). InnerVolumeSpecName "kube-api-access-st6rb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.559419 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-scripts" (OuterVolumeSpecName: "scripts") pod "f8be859d-3be2-4c3f-b10e-b0d04d8fe14d" (UID: "f8be859d-3be2-4c3f-b10e-b0d04d8fe14d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.570508 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.570507 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.589302 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8be859d-3be2-4c3f-b10e-b0d04d8fe14d" (UID: "f8be859d-3be2-4c3f-b10e-b0d04d8fe14d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.609103 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-config-data" (OuterVolumeSpecName: "config-data") pod "f8be859d-3be2-4c3f-b10e-b0d04d8fe14d" (UID: "f8be859d-3be2-4c3f-b10e-b0d04d8fe14d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.655251 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-config-data\") pod \"nova-metadata-0\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.655313 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvx9r\" (UniqueName: \"kubernetes.io/projected/2c551967-7f0d-4816-830e-eaa0add0d53d-kube-api-access-nvx9r\") pod \"nova-metadata-0\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.655396 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c551967-7f0d-4816-830e-eaa0add0d53d-logs\") pod \"nova-metadata-0\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.655420 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.655542 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.655613 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.655629 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.655645 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st6rb\" (UniqueName: \"kubernetes.io/projected/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-kube-api-access-st6rb\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.655658 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.756698 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.756988 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-config-data\") pod \"nova-metadata-0\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.757012 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvx9r\" (UniqueName: \"kubernetes.io/projected/2c551967-7f0d-4816-830e-eaa0add0d53d-kube-api-access-nvx9r\") pod \"nova-metadata-0\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.757064 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c551967-7f0d-4816-830e-eaa0add0d53d-logs\") pod \"nova-metadata-0\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.757088 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.757885 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c551967-7f0d-4816-830e-eaa0add0d53d-logs\") pod \"nova-metadata-0\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.763804 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.763904 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-config-data\") pod \"nova-metadata-0\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.765231 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.775146 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvx9r\" (UniqueName: \"kubernetes.io/projected/2c551967-7f0d-4816-830e-eaa0add0d53d-kube-api-access-nvx9r\") pod \"nova-metadata-0\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " pod="openstack/nova-metadata-0" Feb 27 16:45:46 crc kubenswrapper[4814]: I0227 16:45:46.999401 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.200060 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qm8cq" event={"ID":"f8be859d-3be2-4c3f-b10e-b0d04d8fe14d","Type":"ContainerDied","Data":"19476a39fed26e2f40be51c86e7d4895cf817cc64dee97e8b19e89c282a644d7"} Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.200105 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19476a39fed26e2f40be51c86e7d4895cf817cc64dee97e8b19e89c282a644d7" Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.200105 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qm8cq" Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.200230 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c3108e0c-07ff-496a-9536-caf85d5dbaba" containerName="nova-scheduler-scheduler" containerID="cri-o://17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a" gracePeriod=30 Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.301097 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.302797 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.309766 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.330324 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.471143 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bac854e-cfa6-4eaa-9272-5ecc27c88d5c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2bac854e-cfa6-4eaa-9272-5ecc27c88d5c\") " pod="openstack/nova-cell1-conductor-0" Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.471243 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bac854e-cfa6-4eaa-9272-5ecc27c88d5c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2bac854e-cfa6-4eaa-9272-5ecc27c88d5c\") " pod="openstack/nova-cell1-conductor-0" Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.471305 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnbv6\" (UniqueName: \"kubernetes.io/projected/2bac854e-cfa6-4eaa-9272-5ecc27c88d5c-kube-api-access-xnbv6\") pod \"nova-cell1-conductor-0\" (UID: \"2bac854e-cfa6-4eaa-9272-5ecc27c88d5c\") " pod="openstack/nova-cell1-conductor-0" Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.537867 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:45:47 crc kubenswrapper[4814]: W0227 16:45:47.544715 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c551967_7f0d_4816_830e_eaa0add0d53d.slice/crio-66fb0c22b33d93cdc644279528e6a01180cdcab2abc193749abea486e45cda06 WatchSource:0}: Error finding container 66fb0c22b33d93cdc644279528e6a01180cdcab2abc193749abea486e45cda06: Status 404 returned error can't find the container with id 66fb0c22b33d93cdc644279528e6a01180cdcab2abc193749abea486e45cda06 Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.574638 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bac854e-cfa6-4eaa-9272-5ecc27c88d5c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2bac854e-cfa6-4eaa-9272-5ecc27c88d5c\") " pod="openstack/nova-cell1-conductor-0" Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.574723 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bac854e-cfa6-4eaa-9272-5ecc27c88d5c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2bac854e-cfa6-4eaa-9272-5ecc27c88d5c\") " pod="openstack/nova-cell1-conductor-0" Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.574817 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnbv6\" (UniqueName: \"kubernetes.io/projected/2bac854e-cfa6-4eaa-9272-5ecc27c88d5c-kube-api-access-xnbv6\") pod \"nova-cell1-conductor-0\" (UID: \"2bac854e-cfa6-4eaa-9272-5ecc27c88d5c\") " pod="openstack/nova-cell1-conductor-0" Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.581436 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bac854e-cfa6-4eaa-9272-5ecc27c88d5c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2bac854e-cfa6-4eaa-9272-5ecc27c88d5c\") " pod="openstack/nova-cell1-conductor-0" Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.587529 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bac854e-cfa6-4eaa-9272-5ecc27c88d5c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2bac854e-cfa6-4eaa-9272-5ecc27c88d5c\") " pod="openstack/nova-cell1-conductor-0" Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.604639 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnbv6\" (UniqueName: \"kubernetes.io/projected/2bac854e-cfa6-4eaa-9272-5ecc27c88d5c-kube-api-access-xnbv6\") pod \"nova-cell1-conductor-0\" (UID: \"2bac854e-cfa6-4eaa-9272-5ecc27c88d5c\") " pod="openstack/nova-cell1-conductor-0" Feb 27 16:45:47 crc kubenswrapper[4814]: I0227 16:45:47.627075 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 27 16:45:48 crc kubenswrapper[4814]: I0227 16:45:48.080810 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 27 16:45:48 crc kubenswrapper[4814]: W0227 16:45:48.085086 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bac854e_cfa6_4eaa_9272_5ecc27c88d5c.slice/crio-7d9a7e3b7b41f2bc35469f0ff40d5af513567488ac6f1cd7f23adc6f59519680 WatchSource:0}: Error finding container 7d9a7e3b7b41f2bc35469f0ff40d5af513567488ac6f1cd7f23adc6f59519680: Status 404 returned error can't find the container with id 7d9a7e3b7b41f2bc35469f0ff40d5af513567488ac6f1cd7f23adc6f59519680 Feb 27 16:45:48 crc kubenswrapper[4814]: I0227 16:45:48.213845 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c551967-7f0d-4816-830e-eaa0add0d53d","Type":"ContainerStarted","Data":"9a9c15afaa904dd28044290f22fbf26fff97402adbd1878b5cc23ab62385d2c6"} Feb 27 16:45:48 crc kubenswrapper[4814]: I0227 16:45:48.214206 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c551967-7f0d-4816-830e-eaa0add0d53d","Type":"ContainerStarted","Data":"b14b242cb5aafc2ba966b84df2e900c247ea82c4399c802061210b3fe6a82cf2"} Feb 27 16:45:48 crc kubenswrapper[4814]: I0227 16:45:48.214223 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c551967-7f0d-4816-830e-eaa0add0d53d","Type":"ContainerStarted","Data":"66fb0c22b33d93cdc644279528e6a01180cdcab2abc193749abea486e45cda06"} Feb 27 16:45:48 crc kubenswrapper[4814]: I0227 16:45:48.217613 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2bac854e-cfa6-4eaa-9272-5ecc27c88d5c","Type":"ContainerStarted","Data":"7d9a7e3b7b41f2bc35469f0ff40d5af513567488ac6f1cd7f23adc6f59519680"} Feb 27 16:45:48 crc kubenswrapper[4814]: I0227 16:45:48.244808 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.244787826 podStartE2EDuration="2.244787826s" podCreationTimestamp="2026-02-27 16:45:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:45:48.230710745 +0000 UTC m=+1360.683335585" watchObservedRunningTime="2026-02-27 16:45:48.244787826 +0000 UTC m=+1360.697412656" Feb 27 16:45:48 crc kubenswrapper[4814]: I0227 16:45:48.517981 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27249d07-0ac4-49c6-9040-2fb2f22a59fd" path="/var/lib/kubelet/pods/27249d07-0ac4-49c6-9040-2fb2f22a59fd/volumes" Feb 27 16:45:49 crc kubenswrapper[4814]: I0227 16:45:49.250670 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2bac854e-cfa6-4eaa-9272-5ecc27c88d5c","Type":"ContainerStarted","Data":"c14dfba48d85ccbaf9b403634c978c101d8ede453aa5f7fdd7445021399b4001"} Feb 27 16:45:49 crc kubenswrapper[4814]: I0227 16:45:49.250991 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 27 16:45:49 crc kubenswrapper[4814]: I0227 16:45:49.277836 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.277813466 podStartE2EDuration="2.277813466s" podCreationTimestamp="2026-02-27 16:45:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:45:49.274144432 +0000 UTC m=+1361.726769292" watchObservedRunningTime="2026-02-27 16:45:49.277813466 +0000 UTC m=+1361.730438336" Feb 27 16:45:49 crc kubenswrapper[4814]: E0227 16:45:49.369243 4814 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 16:45:49 crc kubenswrapper[4814]: E0227 16:45:49.374433 4814 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 16:45:49 crc kubenswrapper[4814]: E0227 16:45:49.376680 4814 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 16:45:49 crc kubenswrapper[4814]: E0227 16:45:49.376767 4814 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c3108e0c-07ff-496a-9536-caf85d5dbaba" containerName="nova-scheduler-scheduler" Feb 27 16:45:49 crc kubenswrapper[4814]: E0227 16:45:49.854326 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa51609e_188a_4ded_962b_9f9641404afc.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3108e0c_07ff_496a_9536_caf85d5dbaba.slice/crio-17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa51609e_188a_4ded_962b_9f9641404afc.slice/crio-eb7a32b82b3fba9fb45c72f26ffdb9697916b2788f479c2eb70fe921070b3abe\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3108e0c_07ff_496a_9536_caf85d5dbaba.slice/crio-conmon-17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a.scope\": RecentStats: unable to find data in memory cache]" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.197291 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.258234 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wtct\" (UniqueName: \"kubernetes.io/projected/c3108e0c-07ff-496a-9536-caf85d5dbaba-kube-api-access-8wtct\") pod \"c3108e0c-07ff-496a-9536-caf85d5dbaba\" (UID: \"c3108e0c-07ff-496a-9536-caf85d5dbaba\") " Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.258356 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3108e0c-07ff-496a-9536-caf85d5dbaba-config-data\") pod \"c3108e0c-07ff-496a-9536-caf85d5dbaba\" (UID: \"c3108e0c-07ff-496a-9536-caf85d5dbaba\") " Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.258491 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3108e0c-07ff-496a-9536-caf85d5dbaba-combined-ca-bundle\") pod \"c3108e0c-07ff-496a-9536-caf85d5dbaba\" (UID: \"c3108e0c-07ff-496a-9536-caf85d5dbaba\") " Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.284827 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3108e0c-07ff-496a-9536-caf85d5dbaba-kube-api-access-8wtct" (OuterVolumeSpecName: "kube-api-access-8wtct") pod "c3108e0c-07ff-496a-9536-caf85d5dbaba" (UID: "c3108e0c-07ff-496a-9536-caf85d5dbaba"). InnerVolumeSpecName "kube-api-access-8wtct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.292965 4814 generic.go:334] "Generic (PLEG): container finished" podID="c3108e0c-07ff-496a-9536-caf85d5dbaba" containerID="17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a" exitCode=0 Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.293125 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c3108e0c-07ff-496a-9536-caf85d5dbaba","Type":"ContainerDied","Data":"17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a"} Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.293224 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c3108e0c-07ff-496a-9536-caf85d5dbaba","Type":"ContainerDied","Data":"e5f98139b34fe336f002c666d9abbccbea2eb9c6507c52aa051d875a7f86b75c"} Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.293415 4814 scope.go:117] "RemoveContainer" containerID="17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.293147 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.313873 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3108e0c-07ff-496a-9536-caf85d5dbaba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3108e0c-07ff-496a-9536-caf85d5dbaba" (UID: "c3108e0c-07ff-496a-9536-caf85d5dbaba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.327924 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3108e0c-07ff-496a-9536-caf85d5dbaba-config-data" (OuterVolumeSpecName: "config-data") pod "c3108e0c-07ff-496a-9536-caf85d5dbaba" (UID: "c3108e0c-07ff-496a-9536-caf85d5dbaba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.328108 4814 scope.go:117] "RemoveContainer" containerID="17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a" Feb 27 16:45:50 crc kubenswrapper[4814]: E0227 16:45:50.328687 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a\": container with ID starting with 17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a not found: ID does not exist" containerID="17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.328742 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a"} err="failed to get container status \"17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a\": rpc error: code = NotFound desc = could not find container \"17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a\": container with ID starting with 17a01f1648f5875f1bd7d34dd34f96a6c96d03ea692bf7d0995e93a7dbb4b78a not found: ID does not exist" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.360861 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3108e0c-07ff-496a-9536-caf85d5dbaba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.360902 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wtct\" (UniqueName: \"kubernetes.io/projected/c3108e0c-07ff-496a-9536-caf85d5dbaba-kube-api-access-8wtct\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.360919 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3108e0c-07ff-496a-9536-caf85d5dbaba-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.618851 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.626159 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.642884 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 16:45:50 crc kubenswrapper[4814]: E0227 16:45:50.643277 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3108e0c-07ff-496a-9536-caf85d5dbaba" containerName="nova-scheduler-scheduler" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.643294 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3108e0c-07ff-496a-9536-caf85d5dbaba" containerName="nova-scheduler-scheduler" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.643471 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3108e0c-07ff-496a-9536-caf85d5dbaba" containerName="nova-scheduler-scheduler" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.644068 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.648262 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.664598 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.767213 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d184c7-39b0-416a-9d91-652ff9439ed1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a4d184c7-39b0-416a-9d91-652ff9439ed1\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.767493 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msshh\" (UniqueName: \"kubernetes.io/projected/a4d184c7-39b0-416a-9d91-652ff9439ed1-kube-api-access-msshh\") pod \"nova-scheduler-0\" (UID: \"a4d184c7-39b0-416a-9d91-652ff9439ed1\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.767662 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4d184c7-39b0-416a-9d91-652ff9439ed1-config-data\") pod \"nova-scheduler-0\" (UID: \"a4d184c7-39b0-416a-9d91-652ff9439ed1\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.869398 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msshh\" (UniqueName: \"kubernetes.io/projected/a4d184c7-39b0-416a-9d91-652ff9439ed1-kube-api-access-msshh\") pod \"nova-scheduler-0\" (UID: \"a4d184c7-39b0-416a-9d91-652ff9439ed1\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.869558 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4d184c7-39b0-416a-9d91-652ff9439ed1-config-data\") pod \"nova-scheduler-0\" (UID: \"a4d184c7-39b0-416a-9d91-652ff9439ed1\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.869654 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d184c7-39b0-416a-9d91-652ff9439ed1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a4d184c7-39b0-416a-9d91-652ff9439ed1\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.874938 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d184c7-39b0-416a-9d91-652ff9439ed1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a4d184c7-39b0-416a-9d91-652ff9439ed1\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.875606 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4d184c7-39b0-416a-9d91-652ff9439ed1-config-data\") pod \"nova-scheduler-0\" (UID: \"a4d184c7-39b0-416a-9d91-652ff9439ed1\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:50 crc kubenswrapper[4814]: I0227 16:45:50.892244 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msshh\" (UniqueName: \"kubernetes.io/projected/a4d184c7-39b0-416a-9d91-652ff9439ed1-kube-api-access-msshh\") pod \"nova-scheduler-0\" (UID: \"a4d184c7-39b0-416a-9d91-652ff9439ed1\") " pod="openstack/nova-scheduler-0" Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.012221 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.305548 4814 generic.go:334] "Generic (PLEG): container finished" podID="918450af-cc90-4368-acb7-add3e6d4a737" containerID="97899ac9abe6188d7bb13926164f0fe274525899a6ff5367f2c88ff54c05ea41" exitCode=0 Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.305660 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"918450af-cc90-4368-acb7-add3e6d4a737","Type":"ContainerDied","Data":"97899ac9abe6188d7bb13926164f0fe274525899a6ff5367f2c88ff54c05ea41"} Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.383011 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.563567 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 16:45:51 crc kubenswrapper[4814]: W0227 16:45:51.565079 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4d184c7_39b0_416a_9d91_652ff9439ed1.slice/crio-c4ba718b4566d7d1ca89e92d9f10bdf3a7f084103c3e1f332c2200a06a82e897 WatchSource:0}: Error finding container c4ba718b4566d7d1ca89e92d9f10bdf3a7f084103c3e1f332c2200a06a82e897: Status 404 returned error can't find the container with id c4ba718b4566d7d1ca89e92d9f10bdf3a7f084103c3e1f332c2200a06a82e897 Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.582689 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/918450af-cc90-4368-acb7-add3e6d4a737-config-data\") pod \"918450af-cc90-4368-acb7-add3e6d4a737\" (UID: \"918450af-cc90-4368-acb7-add3e6d4a737\") " Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.582940 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/918450af-cc90-4368-acb7-add3e6d4a737-logs\") pod \"918450af-cc90-4368-acb7-add3e6d4a737\" (UID: \"918450af-cc90-4368-acb7-add3e6d4a737\") " Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.582991 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/918450af-cc90-4368-acb7-add3e6d4a737-combined-ca-bundle\") pod \"918450af-cc90-4368-acb7-add3e6d4a737\" (UID: \"918450af-cc90-4368-acb7-add3e6d4a737\") " Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.583150 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbqcc\" (UniqueName: \"kubernetes.io/projected/918450af-cc90-4368-acb7-add3e6d4a737-kube-api-access-wbqcc\") pod \"918450af-cc90-4368-acb7-add3e6d4a737\" (UID: \"918450af-cc90-4368-acb7-add3e6d4a737\") " Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.583497 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/918450af-cc90-4368-acb7-add3e6d4a737-logs" (OuterVolumeSpecName: "logs") pod "918450af-cc90-4368-acb7-add3e6d4a737" (UID: "918450af-cc90-4368-acb7-add3e6d4a737"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.584573 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/918450af-cc90-4368-acb7-add3e6d4a737-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.587965 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/918450af-cc90-4368-acb7-add3e6d4a737-kube-api-access-wbqcc" (OuterVolumeSpecName: "kube-api-access-wbqcc") pod "918450af-cc90-4368-acb7-add3e6d4a737" (UID: "918450af-cc90-4368-acb7-add3e6d4a737"). InnerVolumeSpecName "kube-api-access-wbqcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.611215 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/918450af-cc90-4368-acb7-add3e6d4a737-config-data" (OuterVolumeSpecName: "config-data") pod "918450af-cc90-4368-acb7-add3e6d4a737" (UID: "918450af-cc90-4368-acb7-add3e6d4a737"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.637691 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/918450af-cc90-4368-acb7-add3e6d4a737-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "918450af-cc90-4368-acb7-add3e6d4a737" (UID: "918450af-cc90-4368-acb7-add3e6d4a737"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.688490 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/918450af-cc90-4368-acb7-add3e6d4a737-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.688537 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbqcc\" (UniqueName: \"kubernetes.io/projected/918450af-cc90-4368-acb7-add3e6d4a737-kube-api-access-wbqcc\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:51 crc kubenswrapper[4814]: I0227 16:45:51.688560 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/918450af-cc90-4368-acb7-add3e6d4a737-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:51.999865 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.000308 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.321959 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"918450af-cc90-4368-acb7-add3e6d4a737","Type":"ContainerDied","Data":"ed46c72ce348ba9ba7c3f2fe840740b26fa676e8403ed5053a3ff15ae8536005"} Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.322307 4814 scope.go:117] "RemoveContainer" containerID="97899ac9abe6188d7bb13926164f0fe274525899a6ff5367f2c88ff54c05ea41" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.322106 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.324679 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a4d184c7-39b0-416a-9d91-652ff9439ed1","Type":"ContainerStarted","Data":"c81475bbb5e6289174b845d8c5921cfe505d523c58204f61cc7e9521e14c5472"} Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.324787 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a4d184c7-39b0-416a-9d91-652ff9439ed1","Type":"ContainerStarted","Data":"c4ba718b4566d7d1ca89e92d9f10bdf3a7f084103c3e1f332c2200a06a82e897"} Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.361229 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.361207916 podStartE2EDuration="2.361207916s" podCreationTimestamp="2026-02-27 16:45:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:45:52.350814091 +0000 UTC m=+1364.803438911" watchObservedRunningTime="2026-02-27 16:45:52.361207916 +0000 UTC m=+1364.813832766" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.369059 4814 scope.go:117] "RemoveContainer" containerID="17c52fa0571b3a706dba4f2fa1daa148aa265d65a8e3de3357e10d43af9cc9ea" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.389682 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.409777 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.431942 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 27 16:45:52 crc kubenswrapper[4814]: E0227 16:45:52.439937 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="918450af-cc90-4368-acb7-add3e6d4a737" containerName="nova-api-api" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.439980 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="918450af-cc90-4368-acb7-add3e6d4a737" containerName="nova-api-api" Feb 27 16:45:52 crc kubenswrapper[4814]: E0227 16:45:52.440000 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="918450af-cc90-4368-acb7-add3e6d4a737" containerName="nova-api-log" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.440008 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="918450af-cc90-4368-acb7-add3e6d4a737" containerName="nova-api-log" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.440317 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="918450af-cc90-4368-acb7-add3e6d4a737" containerName="nova-api-log" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.440343 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="918450af-cc90-4368-acb7-add3e6d4a737" containerName="nova-api-api" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.441523 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.441624 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.444163 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.500933 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="918450af-cc90-4368-acb7-add3e6d4a737" path="/var/lib/kubelet/pods/918450af-cc90-4368-acb7-add3e6d4a737/volumes" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.501764 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3108e0c-07ff-496a-9536-caf85d5dbaba" path="/var/lib/kubelet/pods/c3108e0c-07ff-496a-9536-caf85d5dbaba/volumes" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.605853 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\") " pod="openstack/nova-api-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.606113 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-logs\") pod \"nova-api-0\" (UID: \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\") " pod="openstack/nova-api-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.606186 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74fjk\" (UniqueName: \"kubernetes.io/projected/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-kube-api-access-74fjk\") pod \"nova-api-0\" (UID: \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\") " pod="openstack/nova-api-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.606216 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-config-data\") pod \"nova-api-0\" (UID: \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\") " pod="openstack/nova-api-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.709103 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-logs\") pod \"nova-api-0\" (UID: \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\") " pod="openstack/nova-api-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.709208 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74fjk\" (UniqueName: \"kubernetes.io/projected/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-kube-api-access-74fjk\") pod \"nova-api-0\" (UID: \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\") " pod="openstack/nova-api-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.709250 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-config-data\") pod \"nova-api-0\" (UID: \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\") " pod="openstack/nova-api-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.709364 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\") " pod="openstack/nova-api-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.710499 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-logs\") pod \"nova-api-0\" (UID: \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\") " pod="openstack/nova-api-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.716731 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\") " pod="openstack/nova-api-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.722401 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-config-data\") pod \"nova-api-0\" (UID: \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\") " pod="openstack/nova-api-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.726793 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74fjk\" (UniqueName: \"kubernetes.io/projected/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-kube-api-access-74fjk\") pod \"nova-api-0\" (UID: \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\") " pod="openstack/nova-api-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.768142 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.908414 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.908695 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.908743 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.909423 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5e8afb4996159ac03d9ce86a713b7b4aae36d378d653cf752342458cdea88224"} pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 16:45:52 crc kubenswrapper[4814]: I0227 16:45:52.909466 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" containerID="cri-o://5e8afb4996159ac03d9ce86a713b7b4aae36d378d653cf752342458cdea88224" gracePeriod=600 Feb 27 16:45:53 crc kubenswrapper[4814]: W0227 16:45:53.290844 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode83901de_9140_47e0_8ff1_3dbd0a5c3a70.slice/crio-117be7ebb6ba3597283afeaf874efe88fd089f919ac7abc7c109eb02f44d7214 WatchSource:0}: Error finding container 117be7ebb6ba3597283afeaf874efe88fd089f919ac7abc7c109eb02f44d7214: Status 404 returned error can't find the container with id 117be7ebb6ba3597283afeaf874efe88fd089f919ac7abc7c109eb02f44d7214 Feb 27 16:45:53 crc kubenswrapper[4814]: I0227 16:45:53.291621 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:45:53 crc kubenswrapper[4814]: I0227 16:45:53.350899 4814 generic.go:334] "Generic (PLEG): container finished" podID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerID="5e8afb4996159ac03d9ce86a713b7b4aae36d378d653cf752342458cdea88224" exitCode=0 Feb 27 16:45:53 crc kubenswrapper[4814]: I0227 16:45:53.351007 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerDied","Data":"5e8afb4996159ac03d9ce86a713b7b4aae36d378d653cf752342458cdea88224"} Feb 27 16:45:53 crc kubenswrapper[4814]: I0227 16:45:53.351076 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"e5df7c66a9ca17c8ac1eb1c76463afbe46a665cbf347524e84122f10413b151c"} Feb 27 16:45:53 crc kubenswrapper[4814]: I0227 16:45:53.351105 4814 scope.go:117] "RemoveContainer" containerID="be2cf4df4208edbaae217431871265acc14c77fe4583a6ab07012775c0d5e5f8" Feb 27 16:45:53 crc kubenswrapper[4814]: I0227 16:45:53.354557 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e83901de-9140-47e0-8ff1-3dbd0a5c3a70","Type":"ContainerStarted","Data":"117be7ebb6ba3597283afeaf874efe88fd089f919ac7abc7c109eb02f44d7214"} Feb 27 16:45:54 crc kubenswrapper[4814]: I0227 16:45:54.368306 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e83901de-9140-47e0-8ff1-3dbd0a5c3a70","Type":"ContainerStarted","Data":"2bd4f6b721fa7b11509f5c07799cf9e8be02dea26cb3a291919cb9652c93e0fb"} Feb 27 16:45:54 crc kubenswrapper[4814]: I0227 16:45:54.368954 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e83901de-9140-47e0-8ff1-3dbd0a5c3a70","Type":"ContainerStarted","Data":"9355d3d8b641d777d297a6735f51489dda97883608998231a503e7bb432a50ea"} Feb 27 16:45:54 crc kubenswrapper[4814]: I0227 16:45:54.396482 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.396457663 podStartE2EDuration="2.396457663s" podCreationTimestamp="2026-02-27 16:45:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:45:54.392134488 +0000 UTC m=+1366.844759348" watchObservedRunningTime="2026-02-27 16:45:54.396457663 +0000 UTC m=+1366.849082523" Feb 27 16:45:56 crc kubenswrapper[4814]: I0227 16:45:56.012682 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 27 16:45:56 crc kubenswrapper[4814]: I0227 16:45:56.260762 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 27 16:45:57 crc kubenswrapper[4814]: I0227 16:45:56.999656 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 27 16:45:57 crc kubenswrapper[4814]: I0227 16:45:57.001819 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 27 16:45:57 crc kubenswrapper[4814]: I0227 16:45:57.657026 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 27 16:45:58 crc kubenswrapper[4814]: I0227 16:45:58.025755 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2c551967-7f0d-4816-830e-eaa0add0d53d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 16:45:58 crc kubenswrapper[4814]: I0227 16:45:58.025831 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2c551967-7f0d-4816-830e-eaa0add0d53d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 16:45:59 crc kubenswrapper[4814]: I0227 16:45:59.938220 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 16:45:59 crc kubenswrapper[4814]: I0227 16:45:59.939132 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="e0ea262c-91ce-44fd-b164-be82e6696e93" containerName="kube-state-metrics" containerID="cri-o://91632a104e79e38caed145351cf91b5ac043517bc37f9849b25e7a6cbb580785" gracePeriod=30 Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.135626 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536846-k7bdk"] Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.137010 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536846-k7bdk" Feb 27 16:46:00 crc kubenswrapper[4814]: E0227 16:46:00.137920 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0ea262c_91ce_44fd_b164_be82e6696e93.slice/crio-conmon-91632a104e79e38caed145351cf91b5ac043517bc37f9849b25e7a6cbb580785.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa51609e_188a_4ded_962b_9f9641404afc.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0ea262c_91ce_44fd_b164_be82e6696e93.slice/crio-91632a104e79e38caed145351cf91b5ac043517bc37f9849b25e7a6cbb580785.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa51609e_188a_4ded_962b_9f9641404afc.slice/crio-eb7a32b82b3fba9fb45c72f26ffdb9697916b2788f479c2eb70fe921070b3abe\": RecentStats: unable to find data in memory cache]" Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.147008 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.147387 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.152172 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.163505 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536846-k7bdk"] Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.188205 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67ldd\" (UniqueName: \"kubernetes.io/projected/eb98ddfc-a6c6-4407-9e0e-8f339bed0f03-kube-api-access-67ldd\") pod \"auto-csr-approver-29536846-k7bdk\" (UID: \"eb98ddfc-a6c6-4407-9e0e-8f339bed0f03\") " pod="openshift-infra/auto-csr-approver-29536846-k7bdk" Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.289574 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67ldd\" (UniqueName: \"kubernetes.io/projected/eb98ddfc-a6c6-4407-9e0e-8f339bed0f03-kube-api-access-67ldd\") pod \"auto-csr-approver-29536846-k7bdk\" (UID: \"eb98ddfc-a6c6-4407-9e0e-8f339bed0f03\") " pod="openshift-infra/auto-csr-approver-29536846-k7bdk" Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.315073 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67ldd\" (UniqueName: \"kubernetes.io/projected/eb98ddfc-a6c6-4407-9e0e-8f339bed0f03-kube-api-access-67ldd\") pod \"auto-csr-approver-29536846-k7bdk\" (UID: \"eb98ddfc-a6c6-4407-9e0e-8f339bed0f03\") " pod="openshift-infra/auto-csr-approver-29536846-k7bdk" Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.435846 4814 generic.go:334] "Generic (PLEG): container finished" podID="e0ea262c-91ce-44fd-b164-be82e6696e93" containerID="91632a104e79e38caed145351cf91b5ac043517bc37f9849b25e7a6cbb580785" exitCode=2 Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.435884 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e0ea262c-91ce-44fd-b164-be82e6696e93","Type":"ContainerDied","Data":"91632a104e79e38caed145351cf91b5ac043517bc37f9849b25e7a6cbb580785"} Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.435917 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e0ea262c-91ce-44fd-b164-be82e6696e93","Type":"ContainerDied","Data":"3276ce6178b3b4ca62756118b2280a92a9d07654f250dbe152af22e69dca4af6"} Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.435929 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3276ce6178b3b4ca62756118b2280a92a9d07654f250dbe152af22e69dca4af6" Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.467729 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536846-k7bdk" Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.470631 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.494946 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnn94\" (UniqueName: \"kubernetes.io/projected/e0ea262c-91ce-44fd-b164-be82e6696e93-kube-api-access-pnn94\") pod \"e0ea262c-91ce-44fd-b164-be82e6696e93\" (UID: \"e0ea262c-91ce-44fd-b164-be82e6696e93\") " Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.501520 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0ea262c-91ce-44fd-b164-be82e6696e93-kube-api-access-pnn94" (OuterVolumeSpecName: "kube-api-access-pnn94") pod "e0ea262c-91ce-44fd-b164-be82e6696e93" (UID: "e0ea262c-91ce-44fd-b164-be82e6696e93"). InnerVolumeSpecName "kube-api-access-pnn94". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.598898 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnn94\" (UniqueName: \"kubernetes.io/projected/e0ea262c-91ce-44fd-b164-be82e6696e93-kube-api-access-pnn94\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:00 crc kubenswrapper[4814]: I0227 16:46:00.957288 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536846-k7bdk"] Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.012783 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.042979 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.444377 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.444375 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536846-k7bdk" event={"ID":"eb98ddfc-a6c6-4407-9e0e-8f339bed0f03","Type":"ContainerStarted","Data":"21d73453361c69409edfc27df3eb8b303d62b2318d0abe3d724c5c553fc7e0c9"} Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.477317 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.494481 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.501101 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.513382 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 16:46:01 crc kubenswrapper[4814]: E0227 16:46:01.516868 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ea262c-91ce-44fd-b164-be82e6696e93" containerName="kube-state-metrics" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.516896 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ea262c-91ce-44fd-b164-be82e6696e93" containerName="kube-state-metrics" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.518627 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0ea262c-91ce-44fd-b164-be82e6696e93" containerName="kube-state-metrics" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.525100 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.529022 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.529240 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.555493 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.618670 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtw4q\" (UniqueName: \"kubernetes.io/projected/971d06e4-1037-4da5-b903-27ddda9e9f00-kube-api-access-dtw4q\") pod \"kube-state-metrics-0\" (UID: \"971d06e4-1037-4da5-b903-27ddda9e9f00\") " pod="openstack/kube-state-metrics-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.618770 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/971d06e4-1037-4da5-b903-27ddda9e9f00-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"971d06e4-1037-4da5-b903-27ddda9e9f00\") " pod="openstack/kube-state-metrics-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.618863 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/971d06e4-1037-4da5-b903-27ddda9e9f00-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"971d06e4-1037-4da5-b903-27ddda9e9f00\") " pod="openstack/kube-state-metrics-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.619093 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/971d06e4-1037-4da5-b903-27ddda9e9f00-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"971d06e4-1037-4da5-b903-27ddda9e9f00\") " pod="openstack/kube-state-metrics-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.721239 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/971d06e4-1037-4da5-b903-27ddda9e9f00-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"971d06e4-1037-4da5-b903-27ddda9e9f00\") " pod="openstack/kube-state-metrics-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.721390 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtw4q\" (UniqueName: \"kubernetes.io/projected/971d06e4-1037-4da5-b903-27ddda9e9f00-kube-api-access-dtw4q\") pod \"kube-state-metrics-0\" (UID: \"971d06e4-1037-4da5-b903-27ddda9e9f00\") " pod="openstack/kube-state-metrics-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.721451 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/971d06e4-1037-4da5-b903-27ddda9e9f00-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"971d06e4-1037-4da5-b903-27ddda9e9f00\") " pod="openstack/kube-state-metrics-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.721502 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/971d06e4-1037-4da5-b903-27ddda9e9f00-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"971d06e4-1037-4da5-b903-27ddda9e9f00\") " pod="openstack/kube-state-metrics-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.732782 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/971d06e4-1037-4da5-b903-27ddda9e9f00-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"971d06e4-1037-4da5-b903-27ddda9e9f00\") " pod="openstack/kube-state-metrics-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.733043 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/971d06e4-1037-4da5-b903-27ddda9e9f00-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"971d06e4-1037-4da5-b903-27ddda9e9f00\") " pod="openstack/kube-state-metrics-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.733494 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/971d06e4-1037-4da5-b903-27ddda9e9f00-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"971d06e4-1037-4da5-b903-27ddda9e9f00\") " pod="openstack/kube-state-metrics-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.737457 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtw4q\" (UniqueName: \"kubernetes.io/projected/971d06e4-1037-4da5-b903-27ddda9e9f00-kube-api-access-dtw4q\") pod \"kube-state-metrics-0\" (UID: \"971d06e4-1037-4da5-b903-27ddda9e9f00\") " pod="openstack/kube-state-metrics-0" Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.806826 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.807084 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerName="ceilometer-central-agent" containerID="cri-o://f2a1a11a333a71952c7354ceba5d5534bfd7cd11d4a5fd92841903e1b80046d6" gracePeriod=30 Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.807727 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerName="proxy-httpd" containerID="cri-o://fbe05b43134f6570f2e5fcb9e50bcaacf6cb9dac269d6014816dbe35e87cd8c1" gracePeriod=30 Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.807779 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerName="sg-core" containerID="cri-o://3c93381780739912b02366f3b4a4ed1ec7a7fa4b551cea3ed06fa6c3d954d8d3" gracePeriod=30 Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.807811 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerName="ceilometer-notification-agent" containerID="cri-o://5d4a58b5c3973b8d8eb818fa89598227387781a513be3d96579efd1b35d9fd0b" gracePeriod=30 Feb 27 16:46:01 crc kubenswrapper[4814]: I0227 16:46:01.860444 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 16:46:02 crc kubenswrapper[4814]: I0227 16:46:02.400326 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 16:46:02 crc kubenswrapper[4814]: W0227 16:46:02.401706 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod971d06e4_1037_4da5_b903_27ddda9e9f00.slice/crio-bcb7591c323315b786bd8729e68993d0e84b0cd14dbc8f831ccf460ecdd6aaa0 WatchSource:0}: Error finding container bcb7591c323315b786bd8729e68993d0e84b0cd14dbc8f831ccf460ecdd6aaa0: Status 404 returned error can't find the container with id bcb7591c323315b786bd8729e68993d0e84b0cd14dbc8f831ccf460ecdd6aaa0 Feb 27 16:46:02 crc kubenswrapper[4814]: I0227 16:46:02.472964 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"971d06e4-1037-4da5-b903-27ddda9e9f00","Type":"ContainerStarted","Data":"bcb7591c323315b786bd8729e68993d0e84b0cd14dbc8f831ccf460ecdd6aaa0"} Feb 27 16:46:02 crc kubenswrapper[4814]: I0227 16:46:02.474763 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536846-k7bdk" event={"ID":"eb98ddfc-a6c6-4407-9e0e-8f339bed0f03","Type":"ContainerStarted","Data":"5979d3004d6574e45657308442d2ca78c23c039809420092a8a73557646b8791"} Feb 27 16:46:02 crc kubenswrapper[4814]: I0227 16:46:02.477973 4814 generic.go:334] "Generic (PLEG): container finished" podID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerID="fbe05b43134f6570f2e5fcb9e50bcaacf6cb9dac269d6014816dbe35e87cd8c1" exitCode=0 Feb 27 16:46:02 crc kubenswrapper[4814]: I0227 16:46:02.478011 4814 generic.go:334] "Generic (PLEG): container finished" podID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerID="3c93381780739912b02366f3b4a4ed1ec7a7fa4b551cea3ed06fa6c3d954d8d3" exitCode=2 Feb 27 16:46:02 crc kubenswrapper[4814]: I0227 16:46:02.478022 4814 generic.go:334] "Generic (PLEG): container finished" podID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerID="f2a1a11a333a71952c7354ceba5d5534bfd7cd11d4a5fd92841903e1b80046d6" exitCode=0 Feb 27 16:46:02 crc kubenswrapper[4814]: I0227 16:46:02.478529 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dd4dce7-515e-41bf-9fef-868960c262d7","Type":"ContainerDied","Data":"fbe05b43134f6570f2e5fcb9e50bcaacf6cb9dac269d6014816dbe35e87cd8c1"} Feb 27 16:46:02 crc kubenswrapper[4814]: I0227 16:46:02.478571 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dd4dce7-515e-41bf-9fef-868960c262d7","Type":"ContainerDied","Data":"3c93381780739912b02366f3b4a4ed1ec7a7fa4b551cea3ed06fa6c3d954d8d3"} Feb 27 16:46:02 crc kubenswrapper[4814]: I0227 16:46:02.478585 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dd4dce7-515e-41bf-9fef-868960c262d7","Type":"ContainerDied","Data":"f2a1a11a333a71952c7354ceba5d5534bfd7cd11d4a5fd92841903e1b80046d6"} Feb 27 16:46:02 crc kubenswrapper[4814]: I0227 16:46:02.497998 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536846-k7bdk" podStartSLOduration=1.585256159 podStartE2EDuration="2.497973765s" podCreationTimestamp="2026-02-27 16:46:00 +0000 UTC" firstStartedPulling="2026-02-27 16:46:00.963603079 +0000 UTC m=+1373.416227899" lastFinishedPulling="2026-02-27 16:46:01.876320675 +0000 UTC m=+1374.328945505" observedRunningTime="2026-02-27 16:46:02.488055235 +0000 UTC m=+1374.940680085" watchObservedRunningTime="2026-02-27 16:46:02.497973765 +0000 UTC m=+1374.950598625" Feb 27 16:46:02 crc kubenswrapper[4814]: I0227 16:46:02.498657 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0ea262c-91ce-44fd-b164-be82e6696e93" path="/var/lib/kubelet/pods/e0ea262c-91ce-44fd-b164-be82e6696e93/volumes" Feb 27 16:46:02 crc kubenswrapper[4814]: I0227 16:46:02.768613 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 16:46:02 crc kubenswrapper[4814]: I0227 16:46:02.769001 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 16:46:03 crc kubenswrapper[4814]: I0227 16:46:03.490668 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"971d06e4-1037-4da5-b903-27ddda9e9f00","Type":"ContainerStarted","Data":"d0ca0cf97093cec7eba07f4c234929ac556a6e04e7dfb204b64405269bcd4672"} Feb 27 16:46:03 crc kubenswrapper[4814]: I0227 16:46:03.492333 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 27 16:46:03 crc kubenswrapper[4814]: I0227 16:46:03.498570 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536846-k7bdk" event={"ID":"eb98ddfc-a6c6-4407-9e0e-8f339bed0f03","Type":"ContainerDied","Data":"5979d3004d6574e45657308442d2ca78c23c039809420092a8a73557646b8791"} Feb 27 16:46:03 crc kubenswrapper[4814]: I0227 16:46:03.498643 4814 generic.go:334] "Generic (PLEG): container finished" podID="eb98ddfc-a6c6-4407-9e0e-8f339bed0f03" containerID="5979d3004d6574e45657308442d2ca78c23c039809420092a8a73557646b8791" exitCode=0 Feb 27 16:46:03 crc kubenswrapper[4814]: I0227 16:46:03.529875 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.097100491 podStartE2EDuration="2.52984702s" podCreationTimestamp="2026-02-27 16:46:01 +0000 UTC" firstStartedPulling="2026-02-27 16:46:02.403941153 +0000 UTC m=+1374.856565993" lastFinishedPulling="2026-02-27 16:46:02.836687692 +0000 UTC m=+1375.289312522" observedRunningTime="2026-02-27 16:46:03.517751131 +0000 UTC m=+1375.970376011" watchObservedRunningTime="2026-02-27 16:46:03.52984702 +0000 UTC m=+1375.982471880" Feb 27 16:46:03 crc kubenswrapper[4814]: I0227 16:46:03.850444 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e83901de-9140-47e0-8ff1-3dbd0a5c3a70" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.204:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 16:46:03 crc kubenswrapper[4814]: I0227 16:46:03.850456 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e83901de-9140-47e0-8ff1-3dbd0a5c3a70" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.204:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.121109 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.169110 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-config-data\") pod \"7dd4dce7-515e-41bf-9fef-868960c262d7\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.169173 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-sg-core-conf-yaml\") pod \"7dd4dce7-515e-41bf-9fef-868960c262d7\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.169232 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8sww\" (UniqueName: \"kubernetes.io/projected/7dd4dce7-515e-41bf-9fef-868960c262d7-kube-api-access-k8sww\") pod \"7dd4dce7-515e-41bf-9fef-868960c262d7\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.169267 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dd4dce7-515e-41bf-9fef-868960c262d7-log-httpd\") pod \"7dd4dce7-515e-41bf-9fef-868960c262d7\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.169326 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dd4dce7-515e-41bf-9fef-868960c262d7-run-httpd\") pod \"7dd4dce7-515e-41bf-9fef-868960c262d7\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.169368 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-scripts\") pod \"7dd4dce7-515e-41bf-9fef-868960c262d7\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.169468 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-combined-ca-bundle\") pod \"7dd4dce7-515e-41bf-9fef-868960c262d7\" (UID: \"7dd4dce7-515e-41bf-9fef-868960c262d7\") " Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.169688 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dd4dce7-515e-41bf-9fef-868960c262d7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7dd4dce7-515e-41bf-9fef-868960c262d7" (UID: "7dd4dce7-515e-41bf-9fef-868960c262d7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.170003 4814 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dd4dce7-515e-41bf-9fef-868960c262d7-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.170170 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dd4dce7-515e-41bf-9fef-868960c262d7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7dd4dce7-515e-41bf-9fef-868960c262d7" (UID: "7dd4dce7-515e-41bf-9fef-868960c262d7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.175410 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dd4dce7-515e-41bf-9fef-868960c262d7-kube-api-access-k8sww" (OuterVolumeSpecName: "kube-api-access-k8sww") pod "7dd4dce7-515e-41bf-9fef-868960c262d7" (UID: "7dd4dce7-515e-41bf-9fef-868960c262d7"). InnerVolumeSpecName "kube-api-access-k8sww". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.192560 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-scripts" (OuterVolumeSpecName: "scripts") pod "7dd4dce7-515e-41bf-9fef-868960c262d7" (UID: "7dd4dce7-515e-41bf-9fef-868960c262d7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.203293 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7dd4dce7-515e-41bf-9fef-868960c262d7" (UID: "7dd4dce7-515e-41bf-9fef-868960c262d7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.269433 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7dd4dce7-515e-41bf-9fef-868960c262d7" (UID: "7dd4dce7-515e-41bf-9fef-868960c262d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.272061 4814 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.272090 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8sww\" (UniqueName: \"kubernetes.io/projected/7dd4dce7-515e-41bf-9fef-868960c262d7-kube-api-access-k8sww\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.272102 4814 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7dd4dce7-515e-41bf-9fef-868960c262d7-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.272111 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.272118 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.311238 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-config-data" (OuterVolumeSpecName: "config-data") pod "7dd4dce7-515e-41bf-9fef-868960c262d7" (UID: "7dd4dce7-515e-41bf-9fef-868960c262d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.373611 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd4dce7-515e-41bf-9fef-868960c262d7-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.513691 4814 generic.go:334] "Generic (PLEG): container finished" podID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerID="5d4a58b5c3973b8d8eb818fa89598227387781a513be3d96579efd1b35d9fd0b" exitCode=0 Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.513764 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.513786 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dd4dce7-515e-41bf-9fef-868960c262d7","Type":"ContainerDied","Data":"5d4a58b5c3973b8d8eb818fa89598227387781a513be3d96579efd1b35d9fd0b"} Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.513842 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7dd4dce7-515e-41bf-9fef-868960c262d7","Type":"ContainerDied","Data":"13935b723ba812af0c8e8b8842a40bebb8304d6afeaf6d01dcd2b2cf4abb8531"} Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.513865 4814 scope.go:117] "RemoveContainer" containerID="fbe05b43134f6570f2e5fcb9e50bcaacf6cb9dac269d6014816dbe35e87cd8c1" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.545490 4814 scope.go:117] "RemoveContainer" containerID="3c93381780739912b02366f3b4a4ed1ec7a7fa4b551cea3ed06fa6c3d954d8d3" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.567016 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.589710 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.611590 4814 scope.go:117] "RemoveContainer" containerID="5d4a58b5c3973b8d8eb818fa89598227387781a513be3d96579efd1b35d9fd0b" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.624589 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:46:04 crc kubenswrapper[4814]: E0227 16:46:04.625298 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerName="ceilometer-central-agent" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.625313 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerName="ceilometer-central-agent" Feb 27 16:46:04 crc kubenswrapper[4814]: E0227 16:46:04.625324 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerName="sg-core" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.625331 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerName="sg-core" Feb 27 16:46:04 crc kubenswrapper[4814]: E0227 16:46:04.625343 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerName="proxy-httpd" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.625349 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerName="proxy-httpd" Feb 27 16:46:04 crc kubenswrapper[4814]: E0227 16:46:04.625367 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerName="ceilometer-notification-agent" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.625375 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerName="ceilometer-notification-agent" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.625557 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerName="sg-core" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.625570 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerName="proxy-httpd" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.625584 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerName="ceilometer-central-agent" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.625597 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" containerName="ceilometer-notification-agent" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.627137 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.629990 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.630187 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.630321 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.636665 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.646335 4814 scope.go:117] "RemoveContainer" containerID="f2a1a11a333a71952c7354ceba5d5534bfd7cd11d4a5fd92841903e1b80046d6" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.673432 4814 scope.go:117] "RemoveContainer" containerID="fbe05b43134f6570f2e5fcb9e50bcaacf6cb9dac269d6014816dbe35e87cd8c1" Feb 27 16:46:04 crc kubenswrapper[4814]: E0227 16:46:04.675447 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbe05b43134f6570f2e5fcb9e50bcaacf6cb9dac269d6014816dbe35e87cd8c1\": container with ID starting with fbe05b43134f6570f2e5fcb9e50bcaacf6cb9dac269d6014816dbe35e87cd8c1 not found: ID does not exist" containerID="fbe05b43134f6570f2e5fcb9e50bcaacf6cb9dac269d6014816dbe35e87cd8c1" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.675483 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbe05b43134f6570f2e5fcb9e50bcaacf6cb9dac269d6014816dbe35e87cd8c1"} err="failed to get container status \"fbe05b43134f6570f2e5fcb9e50bcaacf6cb9dac269d6014816dbe35e87cd8c1\": rpc error: code = NotFound desc = could not find container \"fbe05b43134f6570f2e5fcb9e50bcaacf6cb9dac269d6014816dbe35e87cd8c1\": container with ID starting with fbe05b43134f6570f2e5fcb9e50bcaacf6cb9dac269d6014816dbe35e87cd8c1 not found: ID does not exist" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.675509 4814 scope.go:117] "RemoveContainer" containerID="3c93381780739912b02366f3b4a4ed1ec7a7fa4b551cea3ed06fa6c3d954d8d3" Feb 27 16:46:04 crc kubenswrapper[4814]: E0227 16:46:04.675908 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c93381780739912b02366f3b4a4ed1ec7a7fa4b551cea3ed06fa6c3d954d8d3\": container with ID starting with 3c93381780739912b02366f3b4a4ed1ec7a7fa4b551cea3ed06fa6c3d954d8d3 not found: ID does not exist" containerID="3c93381780739912b02366f3b4a4ed1ec7a7fa4b551cea3ed06fa6c3d954d8d3" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.675938 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c93381780739912b02366f3b4a4ed1ec7a7fa4b551cea3ed06fa6c3d954d8d3"} err="failed to get container status \"3c93381780739912b02366f3b4a4ed1ec7a7fa4b551cea3ed06fa6c3d954d8d3\": rpc error: code = NotFound desc = could not find container \"3c93381780739912b02366f3b4a4ed1ec7a7fa4b551cea3ed06fa6c3d954d8d3\": container with ID starting with 3c93381780739912b02366f3b4a4ed1ec7a7fa4b551cea3ed06fa6c3d954d8d3 not found: ID does not exist" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.675961 4814 scope.go:117] "RemoveContainer" containerID="5d4a58b5c3973b8d8eb818fa89598227387781a513be3d96579efd1b35d9fd0b" Feb 27 16:46:04 crc kubenswrapper[4814]: E0227 16:46:04.676185 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d4a58b5c3973b8d8eb818fa89598227387781a513be3d96579efd1b35d9fd0b\": container with ID starting with 5d4a58b5c3973b8d8eb818fa89598227387781a513be3d96579efd1b35d9fd0b not found: ID does not exist" containerID="5d4a58b5c3973b8d8eb818fa89598227387781a513be3d96579efd1b35d9fd0b" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.676204 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d4a58b5c3973b8d8eb818fa89598227387781a513be3d96579efd1b35d9fd0b"} err="failed to get container status \"5d4a58b5c3973b8d8eb818fa89598227387781a513be3d96579efd1b35d9fd0b\": rpc error: code = NotFound desc = could not find container \"5d4a58b5c3973b8d8eb818fa89598227387781a513be3d96579efd1b35d9fd0b\": container with ID starting with 5d4a58b5c3973b8d8eb818fa89598227387781a513be3d96579efd1b35d9fd0b not found: ID does not exist" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.676217 4814 scope.go:117] "RemoveContainer" containerID="f2a1a11a333a71952c7354ceba5d5534bfd7cd11d4a5fd92841903e1b80046d6" Feb 27 16:46:04 crc kubenswrapper[4814]: E0227 16:46:04.676471 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2a1a11a333a71952c7354ceba5d5534bfd7cd11d4a5fd92841903e1b80046d6\": container with ID starting with f2a1a11a333a71952c7354ceba5d5534bfd7cd11d4a5fd92841903e1b80046d6 not found: ID does not exist" containerID="f2a1a11a333a71952c7354ceba5d5534bfd7cd11d4a5fd92841903e1b80046d6" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.676492 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2a1a11a333a71952c7354ceba5d5534bfd7cd11d4a5fd92841903e1b80046d6"} err="failed to get container status \"f2a1a11a333a71952c7354ceba5d5534bfd7cd11d4a5fd92841903e1b80046d6\": rpc error: code = NotFound desc = could not find container \"f2a1a11a333a71952c7354ceba5d5534bfd7cd11d4a5fd92841903e1b80046d6\": container with ID starting with f2a1a11a333a71952c7354ceba5d5534bfd7cd11d4a5fd92841903e1b80046d6 not found: ID does not exist" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.686950 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjl8t\" (UniqueName: \"kubernetes.io/projected/ca9a3048-9e94-4302-8cf4-f69926445fce-kube-api-access-bjl8t\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.687043 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.687067 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca9a3048-9e94-4302-8cf4-f69926445fce-run-httpd\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.687082 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca9a3048-9e94-4302-8cf4-f69926445fce-log-httpd\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.687109 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-config-data\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.687135 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.687368 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-scripts\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.687800 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.790426 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjl8t\" (UniqueName: \"kubernetes.io/projected/ca9a3048-9e94-4302-8cf4-f69926445fce-kube-api-access-bjl8t\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.790584 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.790639 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca9a3048-9e94-4302-8cf4-f69926445fce-run-httpd\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.790681 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca9a3048-9e94-4302-8cf4-f69926445fce-log-httpd\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.790963 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-config-data\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.791026 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.791519 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-scripts\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.791704 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.792694 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca9a3048-9e94-4302-8cf4-f69926445fce-run-httpd\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.792828 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca9a3048-9e94-4302-8cf4-f69926445fce-log-httpd\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.797462 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.797540 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-scripts\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.800034 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-config-data\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.801919 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.803907 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.816528 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjl8t\" (UniqueName: \"kubernetes.io/projected/ca9a3048-9e94-4302-8cf4-f69926445fce-kube-api-access-bjl8t\") pod \"ceilometer-0\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.837110 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536846-k7bdk" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.892734 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67ldd\" (UniqueName: \"kubernetes.io/projected/eb98ddfc-a6c6-4407-9e0e-8f339bed0f03-kube-api-access-67ldd\") pod \"eb98ddfc-a6c6-4407-9e0e-8f339bed0f03\" (UID: \"eb98ddfc-a6c6-4407-9e0e-8f339bed0f03\") " Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.896122 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb98ddfc-a6c6-4407-9e0e-8f339bed0f03-kube-api-access-67ldd" (OuterVolumeSpecName: "kube-api-access-67ldd") pod "eb98ddfc-a6c6-4407-9e0e-8f339bed0f03" (UID: "eb98ddfc-a6c6-4407-9e0e-8f339bed0f03"). InnerVolumeSpecName "kube-api-access-67ldd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.944806 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:46:04 crc kubenswrapper[4814]: I0227 16:46:04.995559 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67ldd\" (UniqueName: \"kubernetes.io/projected/eb98ddfc-a6c6-4407-9e0e-8f339bed0f03-kube-api-access-67ldd\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:05 crc kubenswrapper[4814]: I0227 16:46:05.410027 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:46:05 crc kubenswrapper[4814]: I0227 16:46:05.531428 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca9a3048-9e94-4302-8cf4-f69926445fce","Type":"ContainerStarted","Data":"a4eddcd052d12c1e6dc5057a0deee337aaa4ed1a2354471d0949115c213339dd"} Feb 27 16:46:05 crc kubenswrapper[4814]: I0227 16:46:05.533501 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536846-k7bdk" event={"ID":"eb98ddfc-a6c6-4407-9e0e-8f339bed0f03","Type":"ContainerDied","Data":"21d73453361c69409edfc27df3eb8b303d62b2318d0abe3d724c5c553fc7e0c9"} Feb 27 16:46:05 crc kubenswrapper[4814]: I0227 16:46:05.533562 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21d73453361c69409edfc27df3eb8b303d62b2318d0abe3d724c5c553fc7e0c9" Feb 27 16:46:05 crc kubenswrapper[4814]: I0227 16:46:05.533527 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536846-k7bdk" Feb 27 16:46:05 crc kubenswrapper[4814]: I0227 16:46:05.573758 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536840-rgdnq"] Feb 27 16:46:05 crc kubenswrapper[4814]: I0227 16:46:05.581132 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536840-rgdnq"] Feb 27 16:46:06 crc kubenswrapper[4814]: I0227 16:46:06.505821 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="280aeb00-70d0-4981-a5ee-430bf4014055" path="/var/lib/kubelet/pods/280aeb00-70d0-4981-a5ee-430bf4014055/volumes" Feb 27 16:46:06 crc kubenswrapper[4814]: I0227 16:46:06.507059 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dd4dce7-515e-41bf-9fef-868960c262d7" path="/var/lib/kubelet/pods/7dd4dce7-515e-41bf-9fef-868960c262d7/volumes" Feb 27 16:46:06 crc kubenswrapper[4814]: I0227 16:46:06.545085 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca9a3048-9e94-4302-8cf4-f69926445fce","Type":"ContainerStarted","Data":"ae11488ce6842204614239a1749df54cb60f3f3f4c76084534137393ef1d704d"} Feb 27 16:46:07 crc kubenswrapper[4814]: I0227 16:46:07.007824 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 27 16:46:07 crc kubenswrapper[4814]: I0227 16:46:07.020068 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 27 16:46:07 crc kubenswrapper[4814]: I0227 16:46:07.023678 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 27 16:46:07 crc kubenswrapper[4814]: I0227 16:46:07.563447 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca9a3048-9e94-4302-8cf4-f69926445fce","Type":"ContainerStarted","Data":"73dfd4efc723d190c509af29444bf99b19bed320fc608837b47bf2423296d540"} Feb 27 16:46:07 crc kubenswrapper[4814]: I0227 16:46:07.569497 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 27 16:46:08 crc kubenswrapper[4814]: I0227 16:46:08.571569 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca9a3048-9e94-4302-8cf4-f69926445fce","Type":"ContainerStarted","Data":"2a3e97aee4608c3001613dd9011a9b324743db6d179ad52654f1490a2a1b169e"} Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.559098 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.590505 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca9a3048-9e94-4302-8cf4-f69926445fce","Type":"ContainerStarted","Data":"bdd99b0f8ac9be75892a7c1419e5bdfe0555bf46542f36b03c7d81f40b49e82f"} Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.591374 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.592645 4814 generic.go:334] "Generic (PLEG): container finished" podID="2ebe342f-0c82-435c-97f3-3d32d607dae1" containerID="b7f23a798416a3ef7914d9218aab56b18b95ad2543be35aa29983c847a419d3d" exitCode=137 Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.593115 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2ebe342f-0c82-435c-97f3-3d32d607dae1","Type":"ContainerDied","Data":"b7f23a798416a3ef7914d9218aab56b18b95ad2543be35aa29983c847a419d3d"} Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.593174 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2ebe342f-0c82-435c-97f3-3d32d607dae1","Type":"ContainerDied","Data":"abba409ed86941f8b4ba3cd6d005ebed1563192b18590bbb6dd77cf8a2568467"} Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.593203 4814 scope.go:117] "RemoveContainer" containerID="b7f23a798416a3ef7914d9218aab56b18b95ad2543be35aa29983c847a419d3d" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.593393 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.621318 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9966349380000001 podStartE2EDuration="5.621298303s" podCreationTimestamp="2026-02-27 16:46:04 +0000 UTC" firstStartedPulling="2026-02-27 16:46:05.403464539 +0000 UTC m=+1377.856089379" lastFinishedPulling="2026-02-27 16:46:09.028127884 +0000 UTC m=+1381.480752744" observedRunningTime="2026-02-27 16:46:09.614245591 +0000 UTC m=+1382.066870421" watchObservedRunningTime="2026-02-27 16:46:09.621298303 +0000 UTC m=+1382.073923133" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.630044 4814 scope.go:117] "RemoveContainer" containerID="b7f23a798416a3ef7914d9218aab56b18b95ad2543be35aa29983c847a419d3d" Feb 27 16:46:09 crc kubenswrapper[4814]: E0227 16:46:09.630628 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7f23a798416a3ef7914d9218aab56b18b95ad2543be35aa29983c847a419d3d\": container with ID starting with b7f23a798416a3ef7914d9218aab56b18b95ad2543be35aa29983c847a419d3d not found: ID does not exist" containerID="b7f23a798416a3ef7914d9218aab56b18b95ad2543be35aa29983c847a419d3d" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.630744 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7f23a798416a3ef7914d9218aab56b18b95ad2543be35aa29983c847a419d3d"} err="failed to get container status \"b7f23a798416a3ef7914d9218aab56b18b95ad2543be35aa29983c847a419d3d\": rpc error: code = NotFound desc = could not find container \"b7f23a798416a3ef7914d9218aab56b18b95ad2543be35aa29983c847a419d3d\": container with ID starting with b7f23a798416a3ef7914d9218aab56b18b95ad2543be35aa29983c847a419d3d not found: ID does not exist" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.688713 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ebe342f-0c82-435c-97f3-3d32d607dae1-combined-ca-bundle\") pod \"2ebe342f-0c82-435c-97f3-3d32d607dae1\" (UID: \"2ebe342f-0c82-435c-97f3-3d32d607dae1\") " Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.689141 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ebe342f-0c82-435c-97f3-3d32d607dae1-config-data\") pod \"2ebe342f-0c82-435c-97f3-3d32d607dae1\" (UID: \"2ebe342f-0c82-435c-97f3-3d32d607dae1\") " Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.689803 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7sf8\" (UniqueName: \"kubernetes.io/projected/2ebe342f-0c82-435c-97f3-3d32d607dae1-kube-api-access-v7sf8\") pod \"2ebe342f-0c82-435c-97f3-3d32d607dae1\" (UID: \"2ebe342f-0c82-435c-97f3-3d32d607dae1\") " Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.696090 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ebe342f-0c82-435c-97f3-3d32d607dae1-kube-api-access-v7sf8" (OuterVolumeSpecName: "kube-api-access-v7sf8") pod "2ebe342f-0c82-435c-97f3-3d32d607dae1" (UID: "2ebe342f-0c82-435c-97f3-3d32d607dae1"). InnerVolumeSpecName "kube-api-access-v7sf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.721861 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ebe342f-0c82-435c-97f3-3d32d607dae1-config-data" (OuterVolumeSpecName: "config-data") pod "2ebe342f-0c82-435c-97f3-3d32d607dae1" (UID: "2ebe342f-0c82-435c-97f3-3d32d607dae1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.723024 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ebe342f-0c82-435c-97f3-3d32d607dae1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ebe342f-0c82-435c-97f3-3d32d607dae1" (UID: "2ebe342f-0c82-435c-97f3-3d32d607dae1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.792544 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7sf8\" (UniqueName: \"kubernetes.io/projected/2ebe342f-0c82-435c-97f3-3d32d607dae1-kube-api-access-v7sf8\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.792591 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ebe342f-0c82-435c-97f3-3d32d607dae1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.792605 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ebe342f-0c82-435c-97f3-3d32d607dae1-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.946952 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.963660 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.984590 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 16:46:09 crc kubenswrapper[4814]: E0227 16:46:09.985412 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb98ddfc-a6c6-4407-9e0e-8f339bed0f03" containerName="oc" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.985511 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb98ddfc-a6c6-4407-9e0e-8f339bed0f03" containerName="oc" Feb 27 16:46:09 crc kubenswrapper[4814]: E0227 16:46:09.985613 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ebe342f-0c82-435c-97f3-3d32d607dae1" containerName="nova-cell1-novncproxy-novncproxy" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.985682 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ebe342f-0c82-435c-97f3-3d32d607dae1" containerName="nova-cell1-novncproxy-novncproxy" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.985983 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ebe342f-0c82-435c-97f3-3d32d607dae1" containerName="nova-cell1-novncproxy-novncproxy" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.986081 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb98ddfc-a6c6-4407-9e0e-8f339bed0f03" containerName="oc" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.986927 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.989346 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.991803 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 27 16:46:09 crc kubenswrapper[4814]: I0227 16:46:09.992080 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.050723 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.150717 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4b56f88-0433-4ef4-b4d7-fec94f0f410a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d4b56f88-0433-4ef4-b4d7-fec94f0f410a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.150762 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vzn5\" (UniqueName: \"kubernetes.io/projected/d4b56f88-0433-4ef4-b4d7-fec94f0f410a-kube-api-access-8vzn5\") pod \"nova-cell1-novncproxy-0\" (UID: \"d4b56f88-0433-4ef4-b4d7-fec94f0f410a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.150821 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4b56f88-0433-4ef4-b4d7-fec94f0f410a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d4b56f88-0433-4ef4-b4d7-fec94f0f410a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.150862 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4b56f88-0433-4ef4-b4d7-fec94f0f410a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d4b56f88-0433-4ef4-b4d7-fec94f0f410a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.150891 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4b56f88-0433-4ef4-b4d7-fec94f0f410a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d4b56f88-0433-4ef4-b4d7-fec94f0f410a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.252247 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4b56f88-0433-4ef4-b4d7-fec94f0f410a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d4b56f88-0433-4ef4-b4d7-fec94f0f410a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.252556 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vzn5\" (UniqueName: \"kubernetes.io/projected/d4b56f88-0433-4ef4-b4d7-fec94f0f410a-kube-api-access-8vzn5\") pod \"nova-cell1-novncproxy-0\" (UID: \"d4b56f88-0433-4ef4-b4d7-fec94f0f410a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.252788 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4b56f88-0433-4ef4-b4d7-fec94f0f410a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d4b56f88-0433-4ef4-b4d7-fec94f0f410a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.252899 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4b56f88-0433-4ef4-b4d7-fec94f0f410a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d4b56f88-0433-4ef4-b4d7-fec94f0f410a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.252959 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4b56f88-0433-4ef4-b4d7-fec94f0f410a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d4b56f88-0433-4ef4-b4d7-fec94f0f410a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.258013 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4b56f88-0433-4ef4-b4d7-fec94f0f410a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d4b56f88-0433-4ef4-b4d7-fec94f0f410a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.258684 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4b56f88-0433-4ef4-b4d7-fec94f0f410a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d4b56f88-0433-4ef4-b4d7-fec94f0f410a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.261911 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4b56f88-0433-4ef4-b4d7-fec94f0f410a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d4b56f88-0433-4ef4-b4d7-fec94f0f410a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.271446 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vzn5\" (UniqueName: \"kubernetes.io/projected/d4b56f88-0433-4ef4-b4d7-fec94f0f410a-kube-api-access-8vzn5\") pod \"nova-cell1-novncproxy-0\" (UID: \"d4b56f88-0433-4ef4-b4d7-fec94f0f410a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.284319 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4b56f88-0433-4ef4-b4d7-fec94f0f410a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d4b56f88-0433-4ef4-b4d7-fec94f0f410a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.326397 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.516886 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ebe342f-0c82-435c-97f3-3d32d607dae1" path="/var/lib/kubelet/pods/2ebe342f-0c82-435c-97f3-3d32d607dae1/volumes" Feb 27 16:46:10 crc kubenswrapper[4814]: I0227 16:46:10.804127 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 16:46:10 crc kubenswrapper[4814]: W0227 16:46:10.805642 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4b56f88_0433_4ef4_b4d7_fec94f0f410a.slice/crio-cd3fc87a459c0a945b3e5e2bd952a6523f22d727cc21915c5e89e9eab6aacd89 WatchSource:0}: Error finding container cd3fc87a459c0a945b3e5e2bd952a6523f22d727cc21915c5e89e9eab6aacd89: Status 404 returned error can't find the container with id cd3fc87a459c0a945b3e5e2bd952a6523f22d727cc21915c5e89e9eab6aacd89 Feb 27 16:46:11 crc kubenswrapper[4814]: I0227 16:46:11.612231 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d4b56f88-0433-4ef4-b4d7-fec94f0f410a","Type":"ContainerStarted","Data":"2d03e2da18b6ac16f53ebc7b80e8fecd543929593319ca0863f5594de60302ad"} Feb 27 16:46:11 crc kubenswrapper[4814]: I0227 16:46:11.612669 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d4b56f88-0433-4ef4-b4d7-fec94f0f410a","Type":"ContainerStarted","Data":"cd3fc87a459c0a945b3e5e2bd952a6523f22d727cc21915c5e89e9eab6aacd89"} Feb 27 16:46:11 crc kubenswrapper[4814]: I0227 16:46:11.657086 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.6570656660000003 podStartE2EDuration="2.657065666s" podCreationTimestamp="2026-02-27 16:46:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:46:11.635756498 +0000 UTC m=+1384.088381358" watchObservedRunningTime="2026-02-27 16:46:11.657065666 +0000 UTC m=+1384.109690506" Feb 27 16:46:11 crc kubenswrapper[4814]: I0227 16:46:11.878328 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 27 16:46:12 crc kubenswrapper[4814]: I0227 16:46:12.772691 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 27 16:46:12 crc kubenswrapper[4814]: I0227 16:46:12.773582 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 27 16:46:12 crc kubenswrapper[4814]: I0227 16:46:12.774011 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 27 16:46:12 crc kubenswrapper[4814]: I0227 16:46:12.774077 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 27 16:46:12 crc kubenswrapper[4814]: I0227 16:46:12.776704 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 27 16:46:12 crc kubenswrapper[4814]: I0227 16:46:12.790773 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 27 16:46:12 crc kubenswrapper[4814]: I0227 16:46:12.994845 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-vbvs4"] Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.002550 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.010654 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-vbvs4"] Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.116927 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.117736 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pggk\" (UniqueName: \"kubernetes.io/projected/2bb45fed-87b8-44f9-8afb-659219445949-kube-api-access-5pggk\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.117878 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.117959 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.118042 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.118281 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-config\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.219786 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pggk\" (UniqueName: \"kubernetes.io/projected/2bb45fed-87b8-44f9-8afb-659219445949-kube-api-access-5pggk\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.219846 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.219872 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.219900 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.219938 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-config\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.219980 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.220952 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.220955 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.221054 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.221634 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-config\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.221801 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.268797 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pggk\" (UniqueName: \"kubernetes.io/projected/2bb45fed-87b8-44f9-8afb-659219445949-kube-api-access-5pggk\") pod \"dnsmasq-dns-59cf4bdb65-vbvs4\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.331226 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:13 crc kubenswrapper[4814]: W0227 16:46:13.841908 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bb45fed_87b8_44f9_8afb_659219445949.slice/crio-b1f70f155b1c700d6f1373ab25d36b9e9e9f0c9ac0c5746628ef194365ec25ea WatchSource:0}: Error finding container b1f70f155b1c700d6f1373ab25d36b9e9e9f0c9ac0c5746628ef194365ec25ea: Status 404 returned error can't find the container with id b1f70f155b1c700d6f1373ab25d36b9e9e9f0c9ac0c5746628ef194365ec25ea Feb 27 16:46:13 crc kubenswrapper[4814]: I0227 16:46:13.842388 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-vbvs4"] Feb 27 16:46:14 crc kubenswrapper[4814]: I0227 16:46:14.648684 4814 generic.go:334] "Generic (PLEG): container finished" podID="2bb45fed-87b8-44f9-8afb-659219445949" containerID="10cbc2745d2e87aa081224dadf4b97d4c4d63f41466a76d95c1d6d9b48dffd83" exitCode=0 Feb 27 16:46:14 crc kubenswrapper[4814]: I0227 16:46:14.648930 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" event={"ID":"2bb45fed-87b8-44f9-8afb-659219445949","Type":"ContainerDied","Data":"10cbc2745d2e87aa081224dadf4b97d4c4d63f41466a76d95c1d6d9b48dffd83"} Feb 27 16:46:14 crc kubenswrapper[4814]: I0227 16:46:14.649090 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" event={"ID":"2bb45fed-87b8-44f9-8afb-659219445949","Type":"ContainerStarted","Data":"b1f70f155b1c700d6f1373ab25d36b9e9e9f0c9ac0c5746628ef194365ec25ea"} Feb 27 16:46:15 crc kubenswrapper[4814]: I0227 16:46:15.327388 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:15 crc kubenswrapper[4814]: I0227 16:46:15.428057 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:46:15 crc kubenswrapper[4814]: I0227 16:46:15.660809 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e83901de-9140-47e0-8ff1-3dbd0a5c3a70" containerName="nova-api-log" containerID="cri-o://9355d3d8b641d777d297a6735f51489dda97883608998231a503e7bb432a50ea" gracePeriod=30 Feb 27 16:46:15 crc kubenswrapper[4814]: I0227 16:46:15.661228 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e83901de-9140-47e0-8ff1-3dbd0a5c3a70" containerName="nova-api-api" containerID="cri-o://2bd4f6b721fa7b11509f5c07799cf9e8be02dea26cb3a291919cb9652c93e0fb" gracePeriod=30 Feb 27 16:46:15 crc kubenswrapper[4814]: I0227 16:46:15.661171 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" event={"ID":"2bb45fed-87b8-44f9-8afb-659219445949","Type":"ContainerStarted","Data":"15849b9d9c0da80aec16722483b871cd48f80d72e782fd62e42d89a0c01a4065"} Feb 27 16:46:15 crc kubenswrapper[4814]: I0227 16:46:15.662184 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:15 crc kubenswrapper[4814]: I0227 16:46:15.690614 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" podStartSLOduration=3.690581502 podStartE2EDuration="3.690581502s" podCreationTimestamp="2026-02-27 16:46:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:46:15.689133256 +0000 UTC m=+1388.141758106" watchObservedRunningTime="2026-02-27 16:46:15.690581502 +0000 UTC m=+1388.143206372" Feb 27 16:46:15 crc kubenswrapper[4814]: I0227 16:46:15.800415 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:46:15 crc kubenswrapper[4814]: I0227 16:46:15.800819 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerName="ceilometer-central-agent" containerID="cri-o://ae11488ce6842204614239a1749df54cb60f3f3f4c76084534137393ef1d704d" gracePeriod=30 Feb 27 16:46:15 crc kubenswrapper[4814]: I0227 16:46:15.801001 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerName="proxy-httpd" containerID="cri-o://bdd99b0f8ac9be75892a7c1419e5bdfe0555bf46542f36b03c7d81f40b49e82f" gracePeriod=30 Feb 27 16:46:15 crc kubenswrapper[4814]: I0227 16:46:15.801359 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerName="sg-core" containerID="cri-o://2a3e97aee4608c3001613dd9011a9b324743db6d179ad52654f1490a2a1b169e" gracePeriod=30 Feb 27 16:46:15 crc kubenswrapper[4814]: I0227 16:46:15.801790 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerName="ceilometer-notification-agent" containerID="cri-o://73dfd4efc723d190c509af29444bf99b19bed320fc608837b47bf2423296d540" gracePeriod=30 Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.677077 4814 generic.go:334] "Generic (PLEG): container finished" podID="e83901de-9140-47e0-8ff1-3dbd0a5c3a70" containerID="9355d3d8b641d777d297a6735f51489dda97883608998231a503e7bb432a50ea" exitCode=143 Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.677127 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e83901de-9140-47e0-8ff1-3dbd0a5c3a70","Type":"ContainerDied","Data":"9355d3d8b641d777d297a6735f51489dda97883608998231a503e7bb432a50ea"} Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.680439 4814 generic.go:334] "Generic (PLEG): container finished" podID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerID="bdd99b0f8ac9be75892a7c1419e5bdfe0555bf46542f36b03c7d81f40b49e82f" exitCode=0 Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.680459 4814 generic.go:334] "Generic (PLEG): container finished" podID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerID="2a3e97aee4608c3001613dd9011a9b324743db6d179ad52654f1490a2a1b169e" exitCode=2 Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.680466 4814 generic.go:334] "Generic (PLEG): container finished" podID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerID="73dfd4efc723d190c509af29444bf99b19bed320fc608837b47bf2423296d540" exitCode=0 Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.680472 4814 generic.go:334] "Generic (PLEG): container finished" podID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerID="ae11488ce6842204614239a1749df54cb60f3f3f4c76084534137393ef1d704d" exitCode=0 Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.680517 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca9a3048-9e94-4302-8cf4-f69926445fce","Type":"ContainerDied","Data":"bdd99b0f8ac9be75892a7c1419e5bdfe0555bf46542f36b03c7d81f40b49e82f"} Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.680589 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca9a3048-9e94-4302-8cf4-f69926445fce","Type":"ContainerDied","Data":"2a3e97aee4608c3001613dd9011a9b324743db6d179ad52654f1490a2a1b169e"} Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.680601 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca9a3048-9e94-4302-8cf4-f69926445fce","Type":"ContainerDied","Data":"73dfd4efc723d190c509af29444bf99b19bed320fc608837b47bf2423296d540"} Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.680609 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca9a3048-9e94-4302-8cf4-f69926445fce","Type":"ContainerDied","Data":"ae11488ce6842204614239a1749df54cb60f3f3f4c76084534137393ef1d704d"} Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.751481 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.787908 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca9a3048-9e94-4302-8cf4-f69926445fce-log-httpd\") pod \"ca9a3048-9e94-4302-8cf4-f69926445fce\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.788000 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-sg-core-conf-yaml\") pod \"ca9a3048-9e94-4302-8cf4-f69926445fce\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.788027 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-config-data\") pod \"ca9a3048-9e94-4302-8cf4-f69926445fce\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.788053 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-combined-ca-bundle\") pod \"ca9a3048-9e94-4302-8cf4-f69926445fce\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.788083 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca9a3048-9e94-4302-8cf4-f69926445fce-run-httpd\") pod \"ca9a3048-9e94-4302-8cf4-f69926445fce\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.788100 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-scripts\") pod \"ca9a3048-9e94-4302-8cf4-f69926445fce\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.788123 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjl8t\" (UniqueName: \"kubernetes.io/projected/ca9a3048-9e94-4302-8cf4-f69926445fce-kube-api-access-bjl8t\") pod \"ca9a3048-9e94-4302-8cf4-f69926445fce\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.788154 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-ceilometer-tls-certs\") pod \"ca9a3048-9e94-4302-8cf4-f69926445fce\" (UID: \"ca9a3048-9e94-4302-8cf4-f69926445fce\") " Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.788718 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca9a3048-9e94-4302-8cf4-f69926445fce-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ca9a3048-9e94-4302-8cf4-f69926445fce" (UID: "ca9a3048-9e94-4302-8cf4-f69926445fce"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.789069 4814 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca9a3048-9e94-4302-8cf4-f69926445fce-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.789240 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca9a3048-9e94-4302-8cf4-f69926445fce-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ca9a3048-9e94-4302-8cf4-f69926445fce" (UID: "ca9a3048-9e94-4302-8cf4-f69926445fce"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.794701 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-scripts" (OuterVolumeSpecName: "scripts") pod "ca9a3048-9e94-4302-8cf4-f69926445fce" (UID: "ca9a3048-9e94-4302-8cf4-f69926445fce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.794717 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca9a3048-9e94-4302-8cf4-f69926445fce-kube-api-access-bjl8t" (OuterVolumeSpecName: "kube-api-access-bjl8t") pod "ca9a3048-9e94-4302-8cf4-f69926445fce" (UID: "ca9a3048-9e94-4302-8cf4-f69926445fce"). InnerVolumeSpecName "kube-api-access-bjl8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.821426 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ca9a3048-9e94-4302-8cf4-f69926445fce" (UID: "ca9a3048-9e94-4302-8cf4-f69926445fce"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.859633 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ca9a3048-9e94-4302-8cf4-f69926445fce" (UID: "ca9a3048-9e94-4302-8cf4-f69926445fce"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.890182 4814 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.890211 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.890221 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjl8t\" (UniqueName: \"kubernetes.io/projected/ca9a3048-9e94-4302-8cf4-f69926445fce-kube-api-access-bjl8t\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.890231 4814 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.890240 4814 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca9a3048-9e94-4302-8cf4-f69926445fce-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.897447 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca9a3048-9e94-4302-8cf4-f69926445fce" (UID: "ca9a3048-9e94-4302-8cf4-f69926445fce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.913188 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-config-data" (OuterVolumeSpecName: "config-data") pod "ca9a3048-9e94-4302-8cf4-f69926445fce" (UID: "ca9a3048-9e94-4302-8cf4-f69926445fce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.991313 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:16 crc kubenswrapper[4814]: I0227 16:46:16.991344 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca9a3048-9e94-4302-8cf4-f69926445fce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.697746 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca9a3048-9e94-4302-8cf4-f69926445fce","Type":"ContainerDied","Data":"a4eddcd052d12c1e6dc5057a0deee337aaa4ed1a2354471d0949115c213339dd"} Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.697853 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.698154 4814 scope.go:117] "RemoveContainer" containerID="bdd99b0f8ac9be75892a7c1419e5bdfe0555bf46542f36b03c7d81f40b49e82f" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.736459 4814 scope.go:117] "RemoveContainer" containerID="2a3e97aee4608c3001613dd9011a9b324743db6d179ad52654f1490a2a1b169e" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.756127 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.770492 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.776981 4814 scope.go:117] "RemoveContainer" containerID="73dfd4efc723d190c509af29444bf99b19bed320fc608837b47bf2423296d540" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.781663 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:46:17 crc kubenswrapper[4814]: E0227 16:46:17.782200 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerName="ceilometer-notification-agent" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.782296 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerName="ceilometer-notification-agent" Feb 27 16:46:17 crc kubenswrapper[4814]: E0227 16:46:17.782366 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerName="proxy-httpd" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.782448 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerName="proxy-httpd" Feb 27 16:46:17 crc kubenswrapper[4814]: E0227 16:46:17.782515 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerName="sg-core" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.782576 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerName="sg-core" Feb 27 16:46:17 crc kubenswrapper[4814]: E0227 16:46:17.782680 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerName="ceilometer-central-agent" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.782743 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerName="ceilometer-central-agent" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.783007 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerName="sg-core" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.783098 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerName="ceilometer-notification-agent" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.783174 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerName="proxy-httpd" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.783242 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" containerName="ceilometer-central-agent" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.785497 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.787932 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.788389 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.789552 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.790423 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.831238 4814 scope.go:117] "RemoveContainer" containerID="ae11488ce6842204614239a1749df54cb60f3f3f4c76084534137393ef1d704d" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.909866 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23a14569-9d74-4660-afec-2d91ae3a30a9-scripts\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.909915 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b85j6\" (UniqueName: \"kubernetes.io/projected/23a14569-9d74-4660-afec-2d91ae3a30a9-kube-api-access-b85j6\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.909943 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/23a14569-9d74-4660-afec-2d91ae3a30a9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.910285 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/23a14569-9d74-4660-afec-2d91ae3a30a9-log-httpd\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.910366 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23a14569-9d74-4660-afec-2d91ae3a30a9-config-data\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.910441 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/23a14569-9d74-4660-afec-2d91ae3a30a9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.910800 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/23a14569-9d74-4660-afec-2d91ae3a30a9-run-httpd\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:17 crc kubenswrapper[4814]: I0227 16:46:17.910829 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23a14569-9d74-4660-afec-2d91ae3a30a9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.012952 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/23a14569-9d74-4660-afec-2d91ae3a30a9-run-httpd\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.012988 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23a14569-9d74-4660-afec-2d91ae3a30a9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.013027 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23a14569-9d74-4660-afec-2d91ae3a30a9-scripts\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.013052 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b85j6\" (UniqueName: \"kubernetes.io/projected/23a14569-9d74-4660-afec-2d91ae3a30a9-kube-api-access-b85j6\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.013077 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/23a14569-9d74-4660-afec-2d91ae3a30a9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.013137 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/23a14569-9d74-4660-afec-2d91ae3a30a9-log-httpd\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.013160 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23a14569-9d74-4660-afec-2d91ae3a30a9-config-data\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.013188 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/23a14569-9d74-4660-afec-2d91ae3a30a9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.014222 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/23a14569-9d74-4660-afec-2d91ae3a30a9-log-httpd\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.014416 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/23a14569-9d74-4660-afec-2d91ae3a30a9-run-httpd\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.018802 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/23a14569-9d74-4660-afec-2d91ae3a30a9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.018936 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/23a14569-9d74-4660-afec-2d91ae3a30a9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.019595 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23a14569-9d74-4660-afec-2d91ae3a30a9-scripts\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.020371 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23a14569-9d74-4660-afec-2d91ae3a30a9-config-data\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.021637 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23a14569-9d74-4660-afec-2d91ae3a30a9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.038705 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b85j6\" (UniqueName: \"kubernetes.io/projected/23a14569-9d74-4660-afec-2d91ae3a30a9-kube-api-access-b85j6\") pod \"ceilometer-0\" (UID: \"23a14569-9d74-4660-afec-2d91ae3a30a9\") " pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.122960 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.503896 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca9a3048-9e94-4302-8cf4-f69926445fce" path="/var/lib/kubelet/pods/ca9a3048-9e94-4302-8cf4-f69926445fce/volumes" Feb 27 16:46:18 crc kubenswrapper[4814]: W0227 16:46:18.589434 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23a14569_9d74_4660_afec_2d91ae3a30a9.slice/crio-f900fb4ba3e0cd9aaa04df5ba2dbd936ce1c0961efbb4a1b6a7582d4b7d657d8 WatchSource:0}: Error finding container f900fb4ba3e0cd9aaa04df5ba2dbd936ce1c0961efbb4a1b6a7582d4b7d657d8: Status 404 returned error can't find the container with id f900fb4ba3e0cd9aaa04df5ba2dbd936ce1c0961efbb4a1b6a7582d4b7d657d8 Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.590131 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 16:46:18 crc kubenswrapper[4814]: I0227 16:46:18.713201 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"23a14569-9d74-4660-afec-2d91ae3a30a9","Type":"ContainerStarted","Data":"f900fb4ba3e0cd9aaa04df5ba2dbd936ce1c0961efbb4a1b6a7582d4b7d657d8"} Feb 27 16:46:19 crc kubenswrapper[4814]: I0227 16:46:19.223228 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 16:46:19 crc kubenswrapper[4814]: I0227 16:46:19.245994 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-combined-ca-bundle\") pod \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\" (UID: \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\") " Feb 27 16:46:19 crc kubenswrapper[4814]: I0227 16:46:19.246491 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-logs\") pod \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\" (UID: \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\") " Feb 27 16:46:19 crc kubenswrapper[4814]: I0227 16:46:19.246619 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-config-data\") pod \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\" (UID: \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\") " Feb 27 16:46:19 crc kubenswrapper[4814]: I0227 16:46:19.246666 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74fjk\" (UniqueName: \"kubernetes.io/projected/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-kube-api-access-74fjk\") pod \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\" (UID: \"e83901de-9140-47e0-8ff1-3dbd0a5c3a70\") " Feb 27 16:46:19 crc kubenswrapper[4814]: I0227 16:46:19.247322 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-logs" (OuterVolumeSpecName: "logs") pod "e83901de-9140-47e0-8ff1-3dbd0a5c3a70" (UID: "e83901de-9140-47e0-8ff1-3dbd0a5c3a70"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:46:19 crc kubenswrapper[4814]: I0227 16:46:19.261108 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-kube-api-access-74fjk" (OuterVolumeSpecName: "kube-api-access-74fjk") pod "e83901de-9140-47e0-8ff1-3dbd0a5c3a70" (UID: "e83901de-9140-47e0-8ff1-3dbd0a5c3a70"). InnerVolumeSpecName "kube-api-access-74fjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:46:19 crc kubenswrapper[4814]: I0227 16:46:19.320347 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e83901de-9140-47e0-8ff1-3dbd0a5c3a70" (UID: "e83901de-9140-47e0-8ff1-3dbd0a5c3a70"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:19 crc kubenswrapper[4814]: I0227 16:46:19.331978 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-config-data" (OuterVolumeSpecName: "config-data") pod "e83901de-9140-47e0-8ff1-3dbd0a5c3a70" (UID: "e83901de-9140-47e0-8ff1-3dbd0a5c3a70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:19 crc kubenswrapper[4814]: I0227 16:46:19.348531 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:19 crc kubenswrapper[4814]: I0227 16:46:19.348574 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74fjk\" (UniqueName: \"kubernetes.io/projected/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-kube-api-access-74fjk\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:19 crc kubenswrapper[4814]: I0227 16:46:19.348591 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:19 crc kubenswrapper[4814]: I0227 16:46:19.348602 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e83901de-9140-47e0-8ff1-3dbd0a5c3a70-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.328105 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.335486 4814 generic.go:334] "Generic (PLEG): container finished" podID="e83901de-9140-47e0-8ff1-3dbd0a5c3a70" containerID="2bd4f6b721fa7b11509f5c07799cf9e8be02dea26cb3a291919cb9652c93e0fb" exitCode=0 Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.335573 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e83901de-9140-47e0-8ff1-3dbd0a5c3a70","Type":"ContainerDied","Data":"2bd4f6b721fa7b11509f5c07799cf9e8be02dea26cb3a291919cb9652c93e0fb"} Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.335611 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e83901de-9140-47e0-8ff1-3dbd0a5c3a70","Type":"ContainerDied","Data":"117be7ebb6ba3597283afeaf874efe88fd089f919ac7abc7c109eb02f44d7214"} Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.335638 4814 scope.go:117] "RemoveContainer" containerID="2bd4f6b721fa7b11509f5c07799cf9e8be02dea26cb3a291919cb9652c93e0fb" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.335796 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.343757 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"23a14569-9d74-4660-afec-2d91ae3a30a9","Type":"ContainerStarted","Data":"2f89411a59a9eb1586760134eddd2ae9e73a94342b807e7aeff7efafc4ef9c09"} Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.360777 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.532947 4814 scope.go:117] "RemoveContainer" containerID="9355d3d8b641d777d297a6735f51489dda97883608998231a503e7bb432a50ea" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.553190 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.570177 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.579378 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 27 16:46:20 crc kubenswrapper[4814]: E0227 16:46:20.580908 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e83901de-9140-47e0-8ff1-3dbd0a5c3a70" containerName="nova-api-api" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.580937 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="e83901de-9140-47e0-8ff1-3dbd0a5c3a70" containerName="nova-api-api" Feb 27 16:46:20 crc kubenswrapper[4814]: E0227 16:46:20.580950 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e83901de-9140-47e0-8ff1-3dbd0a5c3a70" containerName="nova-api-log" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.580959 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="e83901de-9140-47e0-8ff1-3dbd0a5c3a70" containerName="nova-api-log" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.581278 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="e83901de-9140-47e0-8ff1-3dbd0a5c3a70" containerName="nova-api-api" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.581316 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="e83901de-9140-47e0-8ff1-3dbd0a5c3a70" containerName="nova-api-log" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.583199 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.602757 4814 scope.go:117] "RemoveContainer" containerID="2bd4f6b721fa7b11509f5c07799cf9e8be02dea26cb3a291919cb9652c93e0fb" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.603222 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.603586 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 27 16:46:20 crc kubenswrapper[4814]: E0227 16:46:20.603783 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bd4f6b721fa7b11509f5c07799cf9e8be02dea26cb3a291919cb9652c93e0fb\": container with ID starting with 2bd4f6b721fa7b11509f5c07799cf9e8be02dea26cb3a291919cb9652c93e0fb not found: ID does not exist" containerID="2bd4f6b721fa7b11509f5c07799cf9e8be02dea26cb3a291919cb9652c93e0fb" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.603845 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bd4f6b721fa7b11509f5c07799cf9e8be02dea26cb3a291919cb9652c93e0fb"} err="failed to get container status \"2bd4f6b721fa7b11509f5c07799cf9e8be02dea26cb3a291919cb9652c93e0fb\": rpc error: code = NotFound desc = could not find container \"2bd4f6b721fa7b11509f5c07799cf9e8be02dea26cb3a291919cb9652c93e0fb\": container with ID starting with 2bd4f6b721fa7b11509f5c07799cf9e8be02dea26cb3a291919cb9652c93e0fb not found: ID does not exist" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.603870 4814 scope.go:117] "RemoveContainer" containerID="9355d3d8b641d777d297a6735f51489dda97883608998231a503e7bb432a50ea" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.603881 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.606609 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:46:20 crc kubenswrapper[4814]: E0227 16:46:20.606898 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9355d3d8b641d777d297a6735f51489dda97883608998231a503e7bb432a50ea\": container with ID starting with 9355d3d8b641d777d297a6735f51489dda97883608998231a503e7bb432a50ea not found: ID does not exist" containerID="9355d3d8b641d777d297a6735f51489dda97883608998231a503e7bb432a50ea" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.606940 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9355d3d8b641d777d297a6735f51489dda97883608998231a503e7bb432a50ea"} err="failed to get container status \"9355d3d8b641d777d297a6735f51489dda97883608998231a503e7bb432a50ea\": rpc error: code = NotFound desc = could not find container \"9355d3d8b641d777d297a6735f51489dda97883608998231a503e7bb432a50ea\": container with ID starting with 9355d3d8b641d777d297a6735f51489dda97883608998231a503e7bb432a50ea not found: ID does not exist" Feb 27 16:46:20 crc kubenswrapper[4814]: E0227 16:46:20.694066 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode83901de_9140_47e0_8ff1_3dbd0a5c3a70.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode83901de_9140_47e0_8ff1_3dbd0a5c3a70.slice/crio-117be7ebb6ba3597283afeaf874efe88fd089f919ac7abc7c109eb02f44d7214\": RecentStats: unable to find data in memory cache]" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.715884 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.715922 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljjtw\" (UniqueName: \"kubernetes.io/projected/5bf82201-1d83-4529-8f88-4de3def59b2b-kube-api-access-ljjtw\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.715944 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-config-data\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.716106 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bf82201-1d83-4529-8f88-4de3def59b2b-logs\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.716373 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.716907 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-public-tls-certs\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.818815 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-public-tls-certs\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.818870 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.818891 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljjtw\" (UniqueName: \"kubernetes.io/projected/5bf82201-1d83-4529-8f88-4de3def59b2b-kube-api-access-ljjtw\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.818911 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-config-data\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.818941 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bf82201-1d83-4529-8f88-4de3def59b2b-logs\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.819007 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.819831 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bf82201-1d83-4529-8f88-4de3def59b2b-logs\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.824910 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.825699 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-config-data\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.833606 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.833871 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-public-tls-certs\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.844785 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljjtw\" (UniqueName: \"kubernetes.io/projected/5bf82201-1d83-4529-8f88-4de3def59b2b-kube-api-access-ljjtw\") pod \"nova-api-0\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " pod="openstack/nova-api-0" Feb 27 16:46:20 crc kubenswrapper[4814]: I0227 16:46:20.923750 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.359117 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"23a14569-9d74-4660-afec-2d91ae3a30a9","Type":"ContainerStarted","Data":"d07926289a86e4054aa9311580cecfd329e71ab5abadd504fe9db5616e8f574f"} Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.385804 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.386569 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.537651 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-l695f"] Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.539365 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-l695f" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.544688 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.544697 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.550458 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-l695f\" (UID: \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\") " pod="openstack/nova-cell1-cell-mapping-l695f" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.550524 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-l695f"] Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.550630 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8zz6\" (UniqueName: \"kubernetes.io/projected/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-kube-api-access-s8zz6\") pod \"nova-cell1-cell-mapping-l695f\" (UID: \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\") " pod="openstack/nova-cell1-cell-mapping-l695f" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.550776 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-scripts\") pod \"nova-cell1-cell-mapping-l695f\" (UID: \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\") " pod="openstack/nova-cell1-cell-mapping-l695f" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.550796 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-config-data\") pod \"nova-cell1-cell-mapping-l695f\" (UID: \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\") " pod="openstack/nova-cell1-cell-mapping-l695f" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.652604 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8zz6\" (UniqueName: \"kubernetes.io/projected/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-kube-api-access-s8zz6\") pod \"nova-cell1-cell-mapping-l695f\" (UID: \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\") " pod="openstack/nova-cell1-cell-mapping-l695f" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.652940 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-scripts\") pod \"nova-cell1-cell-mapping-l695f\" (UID: \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\") " pod="openstack/nova-cell1-cell-mapping-l695f" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.652963 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-config-data\") pod \"nova-cell1-cell-mapping-l695f\" (UID: \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\") " pod="openstack/nova-cell1-cell-mapping-l695f" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.653000 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-l695f\" (UID: \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\") " pod="openstack/nova-cell1-cell-mapping-l695f" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.656994 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-config-data\") pod \"nova-cell1-cell-mapping-l695f\" (UID: \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\") " pod="openstack/nova-cell1-cell-mapping-l695f" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.658351 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-scripts\") pod \"nova-cell1-cell-mapping-l695f\" (UID: \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\") " pod="openstack/nova-cell1-cell-mapping-l695f" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.660542 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-l695f\" (UID: \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\") " pod="openstack/nova-cell1-cell-mapping-l695f" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.669598 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8zz6\" (UniqueName: \"kubernetes.io/projected/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-kube-api-access-s8zz6\") pod \"nova-cell1-cell-mapping-l695f\" (UID: \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\") " pod="openstack/nova-cell1-cell-mapping-l695f" Feb 27 16:46:21 crc kubenswrapper[4814]: I0227 16:46:21.860711 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-l695f" Feb 27 16:46:22 crc kubenswrapper[4814]: I0227 16:46:22.369409 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-l695f"] Feb 27 16:46:22 crc kubenswrapper[4814]: I0227 16:46:22.386324 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5bf82201-1d83-4529-8f88-4de3def59b2b","Type":"ContainerStarted","Data":"3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611"} Feb 27 16:46:22 crc kubenswrapper[4814]: I0227 16:46:22.386367 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5bf82201-1d83-4529-8f88-4de3def59b2b","Type":"ContainerStarted","Data":"290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924"} Feb 27 16:46:22 crc kubenswrapper[4814]: I0227 16:46:22.386381 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5bf82201-1d83-4529-8f88-4de3def59b2b","Type":"ContainerStarted","Data":"84c55c0029066819193c959165b3ddd968aab45916612f9e2637bb7a7783be5e"} Feb 27 16:46:22 crc kubenswrapper[4814]: I0227 16:46:22.397178 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"23a14569-9d74-4660-afec-2d91ae3a30a9","Type":"ContainerStarted","Data":"d2abb59cba69a09c5e77cb3e792b4a863a2c1735bdec4b6e7a2ab63a06f4cb0f"} Feb 27 16:46:22 crc kubenswrapper[4814]: I0227 16:46:22.400838 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-l695f" event={"ID":"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09","Type":"ContainerStarted","Data":"4aca66803d7e9aa2232203eb13a14e1a696fb3a1dfbebfee3364988cac61823e"} Feb 27 16:46:22 crc kubenswrapper[4814]: I0227 16:46:22.413003 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.412982565 podStartE2EDuration="2.412982565s" podCreationTimestamp="2026-02-27 16:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:46:22.406781051 +0000 UTC m=+1394.859405881" watchObservedRunningTime="2026-02-27 16:46:22.412982565 +0000 UTC m=+1394.865607415" Feb 27 16:46:22 crc kubenswrapper[4814]: I0227 16:46:22.505772 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e83901de-9140-47e0-8ff1-3dbd0a5c3a70" path="/var/lib/kubelet/pods/e83901de-9140-47e0-8ff1-3dbd0a5c3a70/volumes" Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.339020 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.416313 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-f5j6n"] Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.416580 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" podUID="4de0189e-b022-42e1-b572-18e584e0d17c" containerName="dnsmasq-dns" containerID="cri-o://0a6fb20559f0d4da93935d261d6ab12128c9cb409086af7e25a93969e2cee4f4" gracePeriod=10 Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.423373 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-l695f" event={"ID":"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09","Type":"ContainerStarted","Data":"a26226c1f8ce87ab249147fa48629d5c0bb6dc47990436b44cfe34d29a5d9ff1"} Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.444844 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-l695f" podStartSLOduration=2.444823218 podStartE2EDuration="2.444823218s" podCreationTimestamp="2026-02-27 16:46:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:46:23.444175698 +0000 UTC m=+1395.896800538" watchObservedRunningTime="2026-02-27 16:46:23.444823218 +0000 UTC m=+1395.897448068" Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.856384 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.901744 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-ovsdbserver-sb\") pod \"4de0189e-b022-42e1-b572-18e584e0d17c\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.901889 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-ovsdbserver-nb\") pod \"4de0189e-b022-42e1-b572-18e584e0d17c\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.901987 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-dns-swift-storage-0\") pod \"4de0189e-b022-42e1-b572-18e584e0d17c\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.902010 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-dns-svc\") pod \"4de0189e-b022-42e1-b572-18e584e0d17c\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.902042 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-config\") pod \"4de0189e-b022-42e1-b572-18e584e0d17c\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.902059 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcl87\" (UniqueName: \"kubernetes.io/projected/4de0189e-b022-42e1-b572-18e584e0d17c-kube-api-access-zcl87\") pod \"4de0189e-b022-42e1-b572-18e584e0d17c\" (UID: \"4de0189e-b022-42e1-b572-18e584e0d17c\") " Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.909548 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4de0189e-b022-42e1-b572-18e584e0d17c-kube-api-access-zcl87" (OuterVolumeSpecName: "kube-api-access-zcl87") pod "4de0189e-b022-42e1-b572-18e584e0d17c" (UID: "4de0189e-b022-42e1-b572-18e584e0d17c"). InnerVolumeSpecName "kube-api-access-zcl87". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.950248 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4de0189e-b022-42e1-b572-18e584e0d17c" (UID: "4de0189e-b022-42e1-b572-18e584e0d17c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.954537 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4de0189e-b022-42e1-b572-18e584e0d17c" (UID: "4de0189e-b022-42e1-b572-18e584e0d17c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.956755 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-config" (OuterVolumeSpecName: "config") pod "4de0189e-b022-42e1-b572-18e584e0d17c" (UID: "4de0189e-b022-42e1-b572-18e584e0d17c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.968631 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4de0189e-b022-42e1-b572-18e584e0d17c" (UID: "4de0189e-b022-42e1-b572-18e584e0d17c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:46:23 crc kubenswrapper[4814]: I0227 16:46:23.986002 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4de0189e-b022-42e1-b572-18e584e0d17c" (UID: "4de0189e-b022-42e1-b572-18e584e0d17c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.004522 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.004545 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.004554 4814 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.004567 4814 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.004575 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4de0189e-b022-42e1-b572-18e584e0d17c-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.004584 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcl87\" (UniqueName: \"kubernetes.io/projected/4de0189e-b022-42e1-b572-18e584e0d17c-kube-api-access-zcl87\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.447824 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"23a14569-9d74-4660-afec-2d91ae3a30a9","Type":"ContainerStarted","Data":"6d0055307f031119329804e7e67c115b0c7a5935eb1e2652f2df548fbe1edced"} Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.448556 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.461221 4814 generic.go:334] "Generic (PLEG): container finished" podID="4de0189e-b022-42e1-b572-18e584e0d17c" containerID="0a6fb20559f0d4da93935d261d6ab12128c9cb409086af7e25a93969e2cee4f4" exitCode=0 Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.462727 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.470771 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" event={"ID":"4de0189e-b022-42e1-b572-18e584e0d17c","Type":"ContainerDied","Data":"0a6fb20559f0d4da93935d261d6ab12128c9cb409086af7e25a93969e2cee4f4"} Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.470872 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-f5j6n" event={"ID":"4de0189e-b022-42e1-b572-18e584e0d17c","Type":"ContainerDied","Data":"3bca4e9419e6621c53311e778eb9d949ceb4eb6bcc0023469f13dd7fbb8e18a3"} Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.470906 4814 scope.go:117] "RemoveContainer" containerID="0a6fb20559f0d4da93935d261d6ab12128c9cb409086af7e25a93969e2cee4f4" Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.529846 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.892142536 podStartE2EDuration="7.529828344s" podCreationTimestamp="2026-02-27 16:46:17 +0000 UTC" firstStartedPulling="2026-02-27 16:46:18.591144012 +0000 UTC m=+1391.043768862" lastFinishedPulling="2026-02-27 16:46:23.22882981 +0000 UTC m=+1395.681454670" observedRunningTime="2026-02-27 16:46:24.511830252 +0000 UTC m=+1396.964455082" watchObservedRunningTime="2026-02-27 16:46:24.529828344 +0000 UTC m=+1396.982453174" Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.546195 4814 scope.go:117] "RemoveContainer" containerID="302cb5ab87f1e8d112c3e97fd57c98544f5a22484607f06843605a6d04860365" Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.567772 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-f5j6n"] Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.599455 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-f5j6n"] Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.601538 4814 scope.go:117] "RemoveContainer" containerID="0a6fb20559f0d4da93935d261d6ab12128c9cb409086af7e25a93969e2cee4f4" Feb 27 16:46:24 crc kubenswrapper[4814]: E0227 16:46:24.616909 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a6fb20559f0d4da93935d261d6ab12128c9cb409086af7e25a93969e2cee4f4\": container with ID starting with 0a6fb20559f0d4da93935d261d6ab12128c9cb409086af7e25a93969e2cee4f4 not found: ID does not exist" containerID="0a6fb20559f0d4da93935d261d6ab12128c9cb409086af7e25a93969e2cee4f4" Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.616957 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a6fb20559f0d4da93935d261d6ab12128c9cb409086af7e25a93969e2cee4f4"} err="failed to get container status \"0a6fb20559f0d4da93935d261d6ab12128c9cb409086af7e25a93969e2cee4f4\": rpc error: code = NotFound desc = could not find container \"0a6fb20559f0d4da93935d261d6ab12128c9cb409086af7e25a93969e2cee4f4\": container with ID starting with 0a6fb20559f0d4da93935d261d6ab12128c9cb409086af7e25a93969e2cee4f4 not found: ID does not exist" Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.616983 4814 scope.go:117] "RemoveContainer" containerID="302cb5ab87f1e8d112c3e97fd57c98544f5a22484607f06843605a6d04860365" Feb 27 16:46:24 crc kubenswrapper[4814]: E0227 16:46:24.618472 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"302cb5ab87f1e8d112c3e97fd57c98544f5a22484607f06843605a6d04860365\": container with ID starting with 302cb5ab87f1e8d112c3e97fd57c98544f5a22484607f06843605a6d04860365 not found: ID does not exist" containerID="302cb5ab87f1e8d112c3e97fd57c98544f5a22484607f06843605a6d04860365" Feb 27 16:46:24 crc kubenswrapper[4814]: I0227 16:46:24.618529 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"302cb5ab87f1e8d112c3e97fd57c98544f5a22484607f06843605a6d04860365"} err="failed to get container status \"302cb5ab87f1e8d112c3e97fd57c98544f5a22484607f06843605a6d04860365\": rpc error: code = NotFound desc = could not find container \"302cb5ab87f1e8d112c3e97fd57c98544f5a22484607f06843605a6d04860365\": container with ID starting with 302cb5ab87f1e8d112c3e97fd57c98544f5a22484607f06843605a6d04860365 not found: ID does not exist" Feb 27 16:46:26 crc kubenswrapper[4814]: I0227 16:46:26.504508 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4de0189e-b022-42e1-b572-18e584e0d17c" path="/var/lib/kubelet/pods/4de0189e-b022-42e1-b572-18e584e0d17c/volumes" Feb 27 16:46:27 crc kubenswrapper[4814]: I0227 16:46:27.499505 4814 generic.go:334] "Generic (PLEG): container finished" podID="6bbf0f83-c41f-427e-b2b4-6dffd46c5d09" containerID="a26226c1f8ce87ab249147fa48629d5c0bb6dc47990436b44cfe34d29a5d9ff1" exitCode=0 Feb 27 16:46:27 crc kubenswrapper[4814]: I0227 16:46:27.499959 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-l695f" event={"ID":"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09","Type":"ContainerDied","Data":"a26226c1f8ce87ab249147fa48629d5c0bb6dc47990436b44cfe34d29a5d9ff1"} Feb 27 16:46:28 crc kubenswrapper[4814]: I0227 16:46:28.985509 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-l695f" Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.008811 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8zz6\" (UniqueName: \"kubernetes.io/projected/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-kube-api-access-s8zz6\") pod \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\" (UID: \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\") " Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.008890 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-combined-ca-bundle\") pod \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\" (UID: \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\") " Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.008964 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-config-data\") pod \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\" (UID: \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\") " Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.009054 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-scripts\") pod \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\" (UID: \"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09\") " Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.018531 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-scripts" (OuterVolumeSpecName: "scripts") pod "6bbf0f83-c41f-427e-b2b4-6dffd46c5d09" (UID: "6bbf0f83-c41f-427e-b2b4-6dffd46c5d09"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.018684 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-kube-api-access-s8zz6" (OuterVolumeSpecName: "kube-api-access-s8zz6") pod "6bbf0f83-c41f-427e-b2b4-6dffd46c5d09" (UID: "6bbf0f83-c41f-427e-b2b4-6dffd46c5d09"). InnerVolumeSpecName "kube-api-access-s8zz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.066079 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6bbf0f83-c41f-427e-b2b4-6dffd46c5d09" (UID: "6bbf0f83-c41f-427e-b2b4-6dffd46c5d09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.067189 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-config-data" (OuterVolumeSpecName: "config-data") pod "6bbf0f83-c41f-427e-b2b4-6dffd46c5d09" (UID: "6bbf0f83-c41f-427e-b2b4-6dffd46c5d09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.112126 4814 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.112189 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8zz6\" (UniqueName: \"kubernetes.io/projected/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-kube-api-access-s8zz6\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.112219 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.112245 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.537949 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-l695f" Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.537724 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-l695f" event={"ID":"6bbf0f83-c41f-427e-b2b4-6dffd46c5d09","Type":"ContainerDied","Data":"4aca66803d7e9aa2232203eb13a14e1a696fb3a1dfbebfee3364988cac61823e"} Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.542328 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4aca66803d7e9aa2232203eb13a14e1a696fb3a1dfbebfee3364988cac61823e" Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.715892 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.716175 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5bf82201-1d83-4529-8f88-4de3def59b2b" containerName="nova-api-log" containerID="cri-o://290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924" gracePeriod=30 Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.716334 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5bf82201-1d83-4529-8f88-4de3def59b2b" containerName="nova-api-api" containerID="cri-o://3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611" gracePeriod=30 Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.812374 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.812980 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2c551967-7f0d-4816-830e-eaa0add0d53d" containerName="nova-metadata-log" containerID="cri-o://b14b242cb5aafc2ba966b84df2e900c247ea82c4399c802061210b3fe6a82cf2" gracePeriod=30 Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.813435 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2c551967-7f0d-4816-830e-eaa0add0d53d" containerName="nova-metadata-metadata" containerID="cri-o://9a9c15afaa904dd28044290f22fbf26fff97402adbd1878b5cc23ab62385d2c6" gracePeriod=30 Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.828776 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 16:46:29 crc kubenswrapper[4814]: I0227 16:46:29.828991 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a4d184c7-39b0-416a-9d91-652ff9439ed1" containerName="nova-scheduler-scheduler" containerID="cri-o://c81475bbb5e6289174b845d8c5921cfe505d523c58204f61cc7e9521e14c5472" gracePeriod=30 Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.381164 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.539434 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljjtw\" (UniqueName: \"kubernetes.io/projected/5bf82201-1d83-4529-8f88-4de3def59b2b-kube-api-access-ljjtw\") pod \"5bf82201-1d83-4529-8f88-4de3def59b2b\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.539506 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-combined-ca-bundle\") pod \"5bf82201-1d83-4529-8f88-4de3def59b2b\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.539596 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-internal-tls-certs\") pod \"5bf82201-1d83-4529-8f88-4de3def59b2b\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.539702 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-config-data\") pod \"5bf82201-1d83-4529-8f88-4de3def59b2b\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.539767 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-public-tls-certs\") pod \"5bf82201-1d83-4529-8f88-4de3def59b2b\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.539853 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bf82201-1d83-4529-8f88-4de3def59b2b-logs\") pod \"5bf82201-1d83-4529-8f88-4de3def59b2b\" (UID: \"5bf82201-1d83-4529-8f88-4de3def59b2b\") " Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.540758 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bf82201-1d83-4529-8f88-4de3def59b2b-logs" (OuterVolumeSpecName: "logs") pod "5bf82201-1d83-4529-8f88-4de3def59b2b" (UID: "5bf82201-1d83-4529-8f88-4de3def59b2b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.547342 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bf82201-1d83-4529-8f88-4de3def59b2b-kube-api-access-ljjtw" (OuterVolumeSpecName: "kube-api-access-ljjtw") pod "5bf82201-1d83-4529-8f88-4de3def59b2b" (UID: "5bf82201-1d83-4529-8f88-4de3def59b2b"). InnerVolumeSpecName "kube-api-access-ljjtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.555065 4814 generic.go:334] "Generic (PLEG): container finished" podID="2c551967-7f0d-4816-830e-eaa0add0d53d" containerID="b14b242cb5aafc2ba966b84df2e900c247ea82c4399c802061210b3fe6a82cf2" exitCode=143 Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.555130 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c551967-7f0d-4816-830e-eaa0add0d53d","Type":"ContainerDied","Data":"b14b242cb5aafc2ba966b84df2e900c247ea82c4399c802061210b3fe6a82cf2"} Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.561440 4814 generic.go:334] "Generic (PLEG): container finished" podID="5bf82201-1d83-4529-8f88-4de3def59b2b" containerID="3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611" exitCode=0 Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.561478 4814 generic.go:334] "Generic (PLEG): container finished" podID="5bf82201-1d83-4529-8f88-4de3def59b2b" containerID="290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924" exitCode=143 Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.561510 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5bf82201-1d83-4529-8f88-4de3def59b2b","Type":"ContainerDied","Data":"3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611"} Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.561524 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.561548 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5bf82201-1d83-4529-8f88-4de3def59b2b","Type":"ContainerDied","Data":"290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924"} Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.561572 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5bf82201-1d83-4529-8f88-4de3def59b2b","Type":"ContainerDied","Data":"84c55c0029066819193c959165b3ddd968aab45916612f9e2637bb7a7783be5e"} Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.561598 4814 scope.go:117] "RemoveContainer" containerID="3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.578430 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-config-data" (OuterVolumeSpecName: "config-data") pod "5bf82201-1d83-4529-8f88-4de3def59b2b" (UID: "5bf82201-1d83-4529-8f88-4de3def59b2b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.578758 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5bf82201-1d83-4529-8f88-4de3def59b2b" (UID: "5bf82201-1d83-4529-8f88-4de3def59b2b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.608939 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5bf82201-1d83-4529-8f88-4de3def59b2b" (UID: "5bf82201-1d83-4529-8f88-4de3def59b2b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.612970 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5bf82201-1d83-4529-8f88-4de3def59b2b" (UID: "5bf82201-1d83-4529-8f88-4de3def59b2b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.641593 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.641624 4814 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.641635 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.641645 4814 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf82201-1d83-4529-8f88-4de3def59b2b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.641659 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bf82201-1d83-4529-8f88-4de3def59b2b-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.641671 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljjtw\" (UniqueName: \"kubernetes.io/projected/5bf82201-1d83-4529-8f88-4de3def59b2b-kube-api-access-ljjtw\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.701608 4814 scope.go:117] "RemoveContainer" containerID="290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.739178 4814 scope.go:117] "RemoveContainer" containerID="3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611" Feb 27 16:46:30 crc kubenswrapper[4814]: E0227 16:46:30.740113 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611\": container with ID starting with 3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611 not found: ID does not exist" containerID="3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.740154 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611"} err="failed to get container status \"3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611\": rpc error: code = NotFound desc = could not find container \"3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611\": container with ID starting with 3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611 not found: ID does not exist" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.740188 4814 scope.go:117] "RemoveContainer" containerID="290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924" Feb 27 16:46:30 crc kubenswrapper[4814]: E0227 16:46:30.742502 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924\": container with ID starting with 290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924 not found: ID does not exist" containerID="290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.742569 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924"} err="failed to get container status \"290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924\": rpc error: code = NotFound desc = could not find container \"290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924\": container with ID starting with 290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924 not found: ID does not exist" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.742609 4814 scope.go:117] "RemoveContainer" containerID="3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.743070 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611"} err="failed to get container status \"3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611\": rpc error: code = NotFound desc = could not find container \"3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611\": container with ID starting with 3808bdb9af65fa0388a444c3b7b6f4cf9b7c9ace1a486ad790203c91bdc8d611 not found: ID does not exist" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.743112 4814 scope.go:117] "RemoveContainer" containerID="290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.743779 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924"} err="failed to get container status \"290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924\": rpc error: code = NotFound desc = could not find container \"290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924\": container with ID starting with 290d9ce19a18eec06b72130b9c378abb57363e496c4224f39d06a0cccb673924 not found: ID does not exist" Feb 27 16:46:30 crc kubenswrapper[4814]: E0227 16:46:30.937889 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bf82201_1d83_4529_8f88_4de3def59b2b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bf82201_1d83_4529_8f88_4de3def59b2b.slice/crio-84c55c0029066819193c959165b3ddd968aab45916612f9e2637bb7a7783be5e\": RecentStats: unable to find data in memory cache]" Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.976175 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:46:30 crc kubenswrapper[4814]: I0227 16:46:30.994593 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.002476 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 27 16:46:31 crc kubenswrapper[4814]: E0227 16:46:31.002854 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bf82201-1d83-4529-8f88-4de3def59b2b" containerName="nova-api-api" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.002881 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bf82201-1d83-4529-8f88-4de3def59b2b" containerName="nova-api-api" Feb 27 16:46:31 crc kubenswrapper[4814]: E0227 16:46:31.002894 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bf82201-1d83-4529-8f88-4de3def59b2b" containerName="nova-api-log" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.002902 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bf82201-1d83-4529-8f88-4de3def59b2b" containerName="nova-api-log" Feb 27 16:46:31 crc kubenswrapper[4814]: E0227 16:46:31.002929 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4de0189e-b022-42e1-b572-18e584e0d17c" containerName="init" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.002936 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4de0189e-b022-42e1-b572-18e584e0d17c" containerName="init" Feb 27 16:46:31 crc kubenswrapper[4814]: E0227 16:46:31.002950 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bbf0f83-c41f-427e-b2b4-6dffd46c5d09" containerName="nova-manage" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.002956 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bbf0f83-c41f-427e-b2b4-6dffd46c5d09" containerName="nova-manage" Feb 27 16:46:31 crc kubenswrapper[4814]: E0227 16:46:31.002967 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4de0189e-b022-42e1-b572-18e584e0d17c" containerName="dnsmasq-dns" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.002974 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4de0189e-b022-42e1-b572-18e584e0d17c" containerName="dnsmasq-dns" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.003147 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bf82201-1d83-4529-8f88-4de3def59b2b" containerName="nova-api-log" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.003159 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bbf0f83-c41f-427e-b2b4-6dffd46c5d09" containerName="nova-manage" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.003169 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4de0189e-b022-42e1-b572-18e584e0d17c" containerName="dnsmasq-dns" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.003178 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bf82201-1d83-4529-8f88-4de3def59b2b" containerName="nova-api-api" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.004070 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.006912 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.007189 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.008107 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.013900 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:46:31 crc kubenswrapper[4814]: E0227 16:46:31.014746 4814 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c81475bbb5e6289174b845d8c5921cfe505d523c58204f61cc7e9521e14c5472" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 16:46:31 crc kubenswrapper[4814]: E0227 16:46:31.018949 4814 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c81475bbb5e6289174b845d8c5921cfe505d523c58204f61cc7e9521e14c5472" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 16:46:31 crc kubenswrapper[4814]: E0227 16:46:31.020123 4814 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c81475bbb5e6289174b845d8c5921cfe505d523c58204f61cc7e9521e14c5472" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 16:46:31 crc kubenswrapper[4814]: E0227 16:46:31.020176 4814 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="a4d184c7-39b0-416a-9d91-652ff9439ed1" containerName="nova-scheduler-scheduler" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.057073 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b31cce6-b000-448f-853e-231f707a2be5-logs\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.057150 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b31cce6-b000-448f-853e-231f707a2be5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.057349 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b31cce6-b000-448f-853e-231f707a2be5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.057441 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b31cce6-b000-448f-853e-231f707a2be5-public-tls-certs\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.057466 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b31cce6-b000-448f-853e-231f707a2be5-config-data\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.057607 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqkpm\" (UniqueName: \"kubernetes.io/projected/1b31cce6-b000-448f-853e-231f707a2be5-kube-api-access-cqkpm\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.159292 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b31cce6-b000-448f-853e-231f707a2be5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.159347 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b31cce6-b000-448f-853e-231f707a2be5-public-tls-certs\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.159367 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b31cce6-b000-448f-853e-231f707a2be5-config-data\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.159397 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqkpm\" (UniqueName: \"kubernetes.io/projected/1b31cce6-b000-448f-853e-231f707a2be5-kube-api-access-cqkpm\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.159462 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b31cce6-b000-448f-853e-231f707a2be5-logs\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.159523 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b31cce6-b000-448f-853e-231f707a2be5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.159944 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b31cce6-b000-448f-853e-231f707a2be5-logs\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.164719 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b31cce6-b000-448f-853e-231f707a2be5-config-data\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.164887 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b31cce6-b000-448f-853e-231f707a2be5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.172836 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b31cce6-b000-448f-853e-231f707a2be5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.174807 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b31cce6-b000-448f-853e-231f707a2be5-public-tls-certs\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.189073 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqkpm\" (UniqueName: \"kubernetes.io/projected/1b31cce6-b000-448f-853e-231f707a2be5-kube-api-access-cqkpm\") pod \"nova-api-0\" (UID: \"1b31cce6-b000-448f-853e-231f707a2be5\") " pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.335552 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 16:46:31 crc kubenswrapper[4814]: I0227 16:46:31.818537 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 16:46:31 crc kubenswrapper[4814]: W0227 16:46:31.821918 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b31cce6_b000_448f_853e_231f707a2be5.slice/crio-0cde5d97bb44eac92c63a93456dc9de7ae1c6f3e675c58065485f21a954aa018 WatchSource:0}: Error finding container 0cde5d97bb44eac92c63a93456dc9de7ae1c6f3e675c58065485f21a954aa018: Status 404 returned error can't find the container with id 0cde5d97bb44eac92c63a93456dc9de7ae1c6f3e675c58065485f21a954aa018 Feb 27 16:46:32 crc kubenswrapper[4814]: I0227 16:46:32.505648 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bf82201-1d83-4529-8f88-4de3def59b2b" path="/var/lib/kubelet/pods/5bf82201-1d83-4529-8f88-4de3def59b2b/volumes" Feb 27 16:46:32 crc kubenswrapper[4814]: I0227 16:46:32.607579 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b31cce6-b000-448f-853e-231f707a2be5","Type":"ContainerStarted","Data":"32394736c366e6f3aeb62135f7a07d39d6be0601b328ccdfc58834a23dd4e634"} Feb 27 16:46:32 crc kubenswrapper[4814]: I0227 16:46:32.607639 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b31cce6-b000-448f-853e-231f707a2be5","Type":"ContainerStarted","Data":"fc0423ac27a05d2319886ebf29a4cba8bd19da61552e2338158f844a0d626d26"} Feb 27 16:46:32 crc kubenswrapper[4814]: I0227 16:46:32.607652 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b31cce6-b000-448f-853e-231f707a2be5","Type":"ContainerStarted","Data":"0cde5d97bb44eac92c63a93456dc9de7ae1c6f3e675c58065485f21a954aa018"} Feb 27 16:46:32 crc kubenswrapper[4814]: I0227 16:46:32.646025 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.645994254 podStartE2EDuration="2.645994254s" podCreationTimestamp="2026-02-27 16:46:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:46:32.634222186 +0000 UTC m=+1405.086847056" watchObservedRunningTime="2026-02-27 16:46:32.645994254 +0000 UTC m=+1405.098619124" Feb 27 16:46:32 crc kubenswrapper[4814]: I0227 16:46:32.955106 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="2c551967-7f0d-4816-830e-eaa0add0d53d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": read tcp 10.217.0.2:47116->10.217.0.201:8775: read: connection reset by peer" Feb 27 16:46:32 crc kubenswrapper[4814]: I0227 16:46:32.955170 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="2c551967-7f0d-4816-830e-eaa0add0d53d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": read tcp 10.217.0.2:47106->10.217.0.201:8775: read: connection reset by peer" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.546513 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.633226 4814 generic.go:334] "Generic (PLEG): container finished" podID="2c551967-7f0d-4816-830e-eaa0add0d53d" containerID="9a9c15afaa904dd28044290f22fbf26fff97402adbd1878b5cc23ab62385d2c6" exitCode=0 Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.633287 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c551967-7f0d-4816-830e-eaa0add0d53d","Type":"ContainerDied","Data":"9a9c15afaa904dd28044290f22fbf26fff97402adbd1878b5cc23ab62385d2c6"} Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.633326 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c551967-7f0d-4816-830e-eaa0add0d53d","Type":"ContainerDied","Data":"66fb0c22b33d93cdc644279528e6a01180cdcab2abc193749abea486e45cda06"} Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.633350 4814 scope.go:117] "RemoveContainer" containerID="9a9c15afaa904dd28044290f22fbf26fff97402adbd1878b5cc23ab62385d2c6" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.633496 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.666339 4814 scope.go:117] "RemoveContainer" containerID="b14b242cb5aafc2ba966b84df2e900c247ea82c4399c802061210b3fe6a82cf2" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.690196 4814 scope.go:117] "RemoveContainer" containerID="9a9c15afaa904dd28044290f22fbf26fff97402adbd1878b5cc23ab62385d2c6" Feb 27 16:46:33 crc kubenswrapper[4814]: E0227 16:46:33.691757 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a9c15afaa904dd28044290f22fbf26fff97402adbd1878b5cc23ab62385d2c6\": container with ID starting with 9a9c15afaa904dd28044290f22fbf26fff97402adbd1878b5cc23ab62385d2c6 not found: ID does not exist" containerID="9a9c15afaa904dd28044290f22fbf26fff97402adbd1878b5cc23ab62385d2c6" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.691788 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a9c15afaa904dd28044290f22fbf26fff97402adbd1878b5cc23ab62385d2c6"} err="failed to get container status \"9a9c15afaa904dd28044290f22fbf26fff97402adbd1878b5cc23ab62385d2c6\": rpc error: code = NotFound desc = could not find container \"9a9c15afaa904dd28044290f22fbf26fff97402adbd1878b5cc23ab62385d2c6\": container with ID starting with 9a9c15afaa904dd28044290f22fbf26fff97402adbd1878b5cc23ab62385d2c6 not found: ID does not exist" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.691808 4814 scope.go:117] "RemoveContainer" containerID="b14b242cb5aafc2ba966b84df2e900c247ea82c4399c802061210b3fe6a82cf2" Feb 27 16:46:33 crc kubenswrapper[4814]: E0227 16:46:33.692113 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b14b242cb5aafc2ba966b84df2e900c247ea82c4399c802061210b3fe6a82cf2\": container with ID starting with b14b242cb5aafc2ba966b84df2e900c247ea82c4399c802061210b3fe6a82cf2 not found: ID does not exist" containerID="b14b242cb5aafc2ba966b84df2e900c247ea82c4399c802061210b3fe6a82cf2" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.692133 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b14b242cb5aafc2ba966b84df2e900c247ea82c4399c802061210b3fe6a82cf2"} err="failed to get container status \"b14b242cb5aafc2ba966b84df2e900c247ea82c4399c802061210b3fe6a82cf2\": rpc error: code = NotFound desc = could not find container \"b14b242cb5aafc2ba966b84df2e900c247ea82c4399c802061210b3fe6a82cf2\": container with ID starting with b14b242cb5aafc2ba966b84df2e900c247ea82c4399c802061210b3fe6a82cf2 not found: ID does not exist" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.720170 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c551967-7f0d-4816-830e-eaa0add0d53d-logs\") pod \"2c551967-7f0d-4816-830e-eaa0add0d53d\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.720312 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-nova-metadata-tls-certs\") pod \"2c551967-7f0d-4816-830e-eaa0add0d53d\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.720361 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-config-data\") pod \"2c551967-7f0d-4816-830e-eaa0add0d53d\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.720518 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-combined-ca-bundle\") pod \"2c551967-7f0d-4816-830e-eaa0add0d53d\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.720673 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c551967-7f0d-4816-830e-eaa0add0d53d-logs" (OuterVolumeSpecName: "logs") pod "2c551967-7f0d-4816-830e-eaa0add0d53d" (UID: "2c551967-7f0d-4816-830e-eaa0add0d53d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.720699 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvx9r\" (UniqueName: \"kubernetes.io/projected/2c551967-7f0d-4816-830e-eaa0add0d53d-kube-api-access-nvx9r\") pod \"2c551967-7f0d-4816-830e-eaa0add0d53d\" (UID: \"2c551967-7f0d-4816-830e-eaa0add0d53d\") " Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.721524 4814 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c551967-7f0d-4816-830e-eaa0add0d53d-logs\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.732277 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c551967-7f0d-4816-830e-eaa0add0d53d-kube-api-access-nvx9r" (OuterVolumeSpecName: "kube-api-access-nvx9r") pod "2c551967-7f0d-4816-830e-eaa0add0d53d" (UID: "2c551967-7f0d-4816-830e-eaa0add0d53d"). InnerVolumeSpecName "kube-api-access-nvx9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.748376 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-config-data" (OuterVolumeSpecName: "config-data") pod "2c551967-7f0d-4816-830e-eaa0add0d53d" (UID: "2c551967-7f0d-4816-830e-eaa0add0d53d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.771207 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c551967-7f0d-4816-830e-eaa0add0d53d" (UID: "2c551967-7f0d-4816-830e-eaa0add0d53d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.791888 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "2c551967-7f0d-4816-830e-eaa0add0d53d" (UID: "2c551967-7f0d-4816-830e-eaa0add0d53d"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.825673 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.825703 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvx9r\" (UniqueName: \"kubernetes.io/projected/2c551967-7f0d-4816-830e-eaa0add0d53d-kube-api-access-nvx9r\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.825715 4814 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.825729 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c551967-7f0d-4816-830e-eaa0add0d53d-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.980315 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:46:33 crc kubenswrapper[4814]: I0227 16:46:33.991206 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.005397 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:46:34 crc kubenswrapper[4814]: E0227 16:46:34.006018 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c551967-7f0d-4816-830e-eaa0add0d53d" containerName="nova-metadata-log" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.006059 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c551967-7f0d-4816-830e-eaa0add0d53d" containerName="nova-metadata-log" Feb 27 16:46:34 crc kubenswrapper[4814]: E0227 16:46:34.006117 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c551967-7f0d-4816-830e-eaa0add0d53d" containerName="nova-metadata-metadata" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.006134 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c551967-7f0d-4816-830e-eaa0add0d53d" containerName="nova-metadata-metadata" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.006496 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c551967-7f0d-4816-830e-eaa0add0d53d" containerName="nova-metadata-metadata" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.006538 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c551967-7f0d-4816-830e-eaa0add0d53d" containerName="nova-metadata-log" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.008225 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.013862 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.014910 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.017244 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.130400 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7902f26-5f7d-4c34-8166-d1857a631858-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b7902f26-5f7d-4c34-8166-d1857a631858\") " pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.130476 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7902f26-5f7d-4c34-8166-d1857a631858-logs\") pod \"nova-metadata-0\" (UID: \"b7902f26-5f7d-4c34-8166-d1857a631858\") " pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.131069 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7902f26-5f7d-4c34-8166-d1857a631858-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b7902f26-5f7d-4c34-8166-d1857a631858\") " pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.131167 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7902f26-5f7d-4c34-8166-d1857a631858-config-data\") pod \"nova-metadata-0\" (UID: \"b7902f26-5f7d-4c34-8166-d1857a631858\") " pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.131298 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck52v\" (UniqueName: \"kubernetes.io/projected/b7902f26-5f7d-4c34-8166-d1857a631858-kube-api-access-ck52v\") pod \"nova-metadata-0\" (UID: \"b7902f26-5f7d-4c34-8166-d1857a631858\") " pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.233234 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7902f26-5f7d-4c34-8166-d1857a631858-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b7902f26-5f7d-4c34-8166-d1857a631858\") " pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.233364 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7902f26-5f7d-4c34-8166-d1857a631858-config-data\") pod \"nova-metadata-0\" (UID: \"b7902f26-5f7d-4c34-8166-d1857a631858\") " pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.233437 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck52v\" (UniqueName: \"kubernetes.io/projected/b7902f26-5f7d-4c34-8166-d1857a631858-kube-api-access-ck52v\") pod \"nova-metadata-0\" (UID: \"b7902f26-5f7d-4c34-8166-d1857a631858\") " pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.233572 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7902f26-5f7d-4c34-8166-d1857a631858-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b7902f26-5f7d-4c34-8166-d1857a631858\") " pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.233626 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7902f26-5f7d-4c34-8166-d1857a631858-logs\") pod \"nova-metadata-0\" (UID: \"b7902f26-5f7d-4c34-8166-d1857a631858\") " pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.234445 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7902f26-5f7d-4c34-8166-d1857a631858-logs\") pod \"nova-metadata-0\" (UID: \"b7902f26-5f7d-4c34-8166-d1857a631858\") " pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.239796 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7902f26-5f7d-4c34-8166-d1857a631858-config-data\") pod \"nova-metadata-0\" (UID: \"b7902f26-5f7d-4c34-8166-d1857a631858\") " pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.244975 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7902f26-5f7d-4c34-8166-d1857a631858-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b7902f26-5f7d-4c34-8166-d1857a631858\") " pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.247055 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7902f26-5f7d-4c34-8166-d1857a631858-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b7902f26-5f7d-4c34-8166-d1857a631858\") " pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.256113 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck52v\" (UniqueName: \"kubernetes.io/projected/b7902f26-5f7d-4c34-8166-d1857a631858-kube-api-access-ck52v\") pod \"nova-metadata-0\" (UID: \"b7902f26-5f7d-4c34-8166-d1857a631858\") " pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.341411 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.503518 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c551967-7f0d-4816-830e-eaa0add0d53d" path="/var/lib/kubelet/pods/2c551967-7f0d-4816-830e-eaa0add0d53d/volumes" Feb 27 16:46:34 crc kubenswrapper[4814]: I0227 16:46:34.896684 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 16:46:34 crc kubenswrapper[4814]: W0227 16:46:34.901538 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7902f26_5f7d_4c34_8166_d1857a631858.slice/crio-4335578c38c2b3a4b1b05d7f8c427856ebd73738492a3feb19a84dd13cf5e5cb WatchSource:0}: Error finding container 4335578c38c2b3a4b1b05d7f8c427856ebd73738492a3feb19a84dd13cf5e5cb: Status 404 returned error can't find the container with id 4335578c38c2b3a4b1b05d7f8c427856ebd73738492a3feb19a84dd13cf5e5cb Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.059324 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.163764 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d184c7-39b0-416a-9d91-652ff9439ed1-combined-ca-bundle\") pod \"a4d184c7-39b0-416a-9d91-652ff9439ed1\" (UID: \"a4d184c7-39b0-416a-9d91-652ff9439ed1\") " Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.163843 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msshh\" (UniqueName: \"kubernetes.io/projected/a4d184c7-39b0-416a-9d91-652ff9439ed1-kube-api-access-msshh\") pod \"a4d184c7-39b0-416a-9d91-652ff9439ed1\" (UID: \"a4d184c7-39b0-416a-9d91-652ff9439ed1\") " Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.163903 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4d184c7-39b0-416a-9d91-652ff9439ed1-config-data\") pod \"a4d184c7-39b0-416a-9d91-652ff9439ed1\" (UID: \"a4d184c7-39b0-416a-9d91-652ff9439ed1\") " Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.168553 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4d184c7-39b0-416a-9d91-652ff9439ed1-kube-api-access-msshh" (OuterVolumeSpecName: "kube-api-access-msshh") pod "a4d184c7-39b0-416a-9d91-652ff9439ed1" (UID: "a4d184c7-39b0-416a-9d91-652ff9439ed1"). InnerVolumeSpecName "kube-api-access-msshh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.193684 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d184c7-39b0-416a-9d91-652ff9439ed1-config-data" (OuterVolumeSpecName: "config-data") pod "a4d184c7-39b0-416a-9d91-652ff9439ed1" (UID: "a4d184c7-39b0-416a-9d91-652ff9439ed1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.212028 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d184c7-39b0-416a-9d91-652ff9439ed1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4d184c7-39b0-416a-9d91-652ff9439ed1" (UID: "a4d184c7-39b0-416a-9d91-652ff9439ed1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.266502 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4d184c7-39b0-416a-9d91-652ff9439ed1-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.266544 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d184c7-39b0-416a-9d91-652ff9439ed1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.266559 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msshh\" (UniqueName: \"kubernetes.io/projected/a4d184c7-39b0-416a-9d91-652ff9439ed1-kube-api-access-msshh\") on node \"crc\" DevicePath \"\"" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.669030 4814 generic.go:334] "Generic (PLEG): container finished" podID="a4d184c7-39b0-416a-9d91-652ff9439ed1" containerID="c81475bbb5e6289174b845d8c5921cfe505d523c58204f61cc7e9521e14c5472" exitCode=0 Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.669097 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.669115 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a4d184c7-39b0-416a-9d91-652ff9439ed1","Type":"ContainerDied","Data":"c81475bbb5e6289174b845d8c5921cfe505d523c58204f61cc7e9521e14c5472"} Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.669515 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a4d184c7-39b0-416a-9d91-652ff9439ed1","Type":"ContainerDied","Data":"c4ba718b4566d7d1ca89e92d9f10bdf3a7f084103c3e1f332c2200a06a82e897"} Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.669538 4814 scope.go:117] "RemoveContainer" containerID="c81475bbb5e6289174b845d8c5921cfe505d523c58204f61cc7e9521e14c5472" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.673097 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7902f26-5f7d-4c34-8166-d1857a631858","Type":"ContainerStarted","Data":"74a70624b110660256ffd6f1dea9bd1646a5708d09a6bf40ec20a82e02bd9925"} Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.673143 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7902f26-5f7d-4c34-8166-d1857a631858","Type":"ContainerStarted","Data":"0b85a649f94a8f0b9f05b6058eb006503ced4ee1409d4630da89f57eeb8e0a6c"} Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.673157 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7902f26-5f7d-4c34-8166-d1857a631858","Type":"ContainerStarted","Data":"4335578c38c2b3a4b1b05d7f8c427856ebd73738492a3feb19a84dd13cf5e5cb"} Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.707885 4814 scope.go:117] "RemoveContainer" containerID="c81475bbb5e6289174b845d8c5921cfe505d523c58204f61cc7e9521e14c5472" Feb 27 16:46:35 crc kubenswrapper[4814]: E0227 16:46:35.708613 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c81475bbb5e6289174b845d8c5921cfe505d523c58204f61cc7e9521e14c5472\": container with ID starting with c81475bbb5e6289174b845d8c5921cfe505d523c58204f61cc7e9521e14c5472 not found: ID does not exist" containerID="c81475bbb5e6289174b845d8c5921cfe505d523c58204f61cc7e9521e14c5472" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.708686 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c81475bbb5e6289174b845d8c5921cfe505d523c58204f61cc7e9521e14c5472"} err="failed to get container status \"c81475bbb5e6289174b845d8c5921cfe505d523c58204f61cc7e9521e14c5472\": rpc error: code = NotFound desc = could not find container \"c81475bbb5e6289174b845d8c5921cfe505d523c58204f61cc7e9521e14c5472\": container with ID starting with c81475bbb5e6289174b845d8c5921cfe505d523c58204f61cc7e9521e14c5472 not found: ID does not exist" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.711146 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.711126964 podStartE2EDuration="2.711126964s" podCreationTimestamp="2026-02-27 16:46:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:46:35.701474651 +0000 UTC m=+1408.154099491" watchObservedRunningTime="2026-02-27 16:46:35.711126964 +0000 UTC m=+1408.163751804" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.731348 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.741231 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.747914 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 16:46:35 crc kubenswrapper[4814]: E0227 16:46:35.748301 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4d184c7-39b0-416a-9d91-652ff9439ed1" containerName="nova-scheduler-scheduler" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.748318 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4d184c7-39b0-416a-9d91-652ff9439ed1" containerName="nova-scheduler-scheduler" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.748490 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4d184c7-39b0-416a-9d91-652ff9439ed1" containerName="nova-scheduler-scheduler" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.749057 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.755214 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.777428 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aace7eb3-3c5c-4290-8ada-b506bd913841-config-data\") pod \"nova-scheduler-0\" (UID: \"aace7eb3-3c5c-4290-8ada-b506bd913841\") " pod="openstack/nova-scheduler-0" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.777783 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj56t\" (UniqueName: \"kubernetes.io/projected/aace7eb3-3c5c-4290-8ada-b506bd913841-kube-api-access-vj56t\") pod \"nova-scheduler-0\" (UID: \"aace7eb3-3c5c-4290-8ada-b506bd913841\") " pod="openstack/nova-scheduler-0" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.777830 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aace7eb3-3c5c-4290-8ada-b506bd913841-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"aace7eb3-3c5c-4290-8ada-b506bd913841\") " pod="openstack/nova-scheduler-0" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.799014 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.880390 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj56t\" (UniqueName: \"kubernetes.io/projected/aace7eb3-3c5c-4290-8ada-b506bd913841-kube-api-access-vj56t\") pod \"nova-scheduler-0\" (UID: \"aace7eb3-3c5c-4290-8ada-b506bd913841\") " pod="openstack/nova-scheduler-0" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.880450 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aace7eb3-3c5c-4290-8ada-b506bd913841-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"aace7eb3-3c5c-4290-8ada-b506bd913841\") " pod="openstack/nova-scheduler-0" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.880573 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aace7eb3-3c5c-4290-8ada-b506bd913841-config-data\") pod \"nova-scheduler-0\" (UID: \"aace7eb3-3c5c-4290-8ada-b506bd913841\") " pod="openstack/nova-scheduler-0" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.885596 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aace7eb3-3c5c-4290-8ada-b506bd913841-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"aace7eb3-3c5c-4290-8ada-b506bd913841\") " pod="openstack/nova-scheduler-0" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.885713 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aace7eb3-3c5c-4290-8ada-b506bd913841-config-data\") pod \"nova-scheduler-0\" (UID: \"aace7eb3-3c5c-4290-8ada-b506bd913841\") " pod="openstack/nova-scheduler-0" Feb 27 16:46:35 crc kubenswrapper[4814]: I0227 16:46:35.905714 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj56t\" (UniqueName: \"kubernetes.io/projected/aace7eb3-3c5c-4290-8ada-b506bd913841-kube-api-access-vj56t\") pod \"nova-scheduler-0\" (UID: \"aace7eb3-3c5c-4290-8ada-b506bd913841\") " pod="openstack/nova-scheduler-0" Feb 27 16:46:36 crc kubenswrapper[4814]: I0227 16:46:36.110493 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 16:46:36 crc kubenswrapper[4814]: I0227 16:46:36.504611 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4d184c7-39b0-416a-9d91-652ff9439ed1" path="/var/lib/kubelet/pods/a4d184c7-39b0-416a-9d91-652ff9439ed1/volumes" Feb 27 16:46:36 crc kubenswrapper[4814]: W0227 16:46:36.642425 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaace7eb3_3c5c_4290_8ada_b506bd913841.slice/crio-b8602e33caa7ed184c0227e60624aa85dccfc226a9e1784797005b22dc477bbc WatchSource:0}: Error finding container b8602e33caa7ed184c0227e60624aa85dccfc226a9e1784797005b22dc477bbc: Status 404 returned error can't find the container with id b8602e33caa7ed184c0227e60624aa85dccfc226a9e1784797005b22dc477bbc Feb 27 16:46:36 crc kubenswrapper[4814]: I0227 16:46:36.644863 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 16:46:36 crc kubenswrapper[4814]: I0227 16:46:36.690000 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"aace7eb3-3c5c-4290-8ada-b506bd913841","Type":"ContainerStarted","Data":"b8602e33caa7ed184c0227e60624aa85dccfc226a9e1784797005b22dc477bbc"} Feb 27 16:46:37 crc kubenswrapper[4814]: I0227 16:46:37.709976 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"aace7eb3-3c5c-4290-8ada-b506bd913841","Type":"ContainerStarted","Data":"15aeff95dad6254895d4b1113aea8fe9e0cae0de23038cbc18a1ba0d5fb2a8a2"} Feb 27 16:46:37 crc kubenswrapper[4814]: I0227 16:46:37.749324 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.7492987810000002 podStartE2EDuration="2.749298781s" podCreationTimestamp="2026-02-27 16:46:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:46:37.739377891 +0000 UTC m=+1410.192002781" watchObservedRunningTime="2026-02-27 16:46:37.749298781 +0000 UTC m=+1410.201923641" Feb 27 16:46:39 crc kubenswrapper[4814]: I0227 16:46:39.341485 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 16:46:39 crc kubenswrapper[4814]: I0227 16:46:39.341842 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 16:46:41 crc kubenswrapper[4814]: I0227 16:46:41.110588 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 27 16:46:41 crc kubenswrapper[4814]: I0227 16:46:41.335899 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 16:46:41 crc kubenswrapper[4814]: I0227 16:46:41.335988 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 16:46:42 crc kubenswrapper[4814]: I0227 16:46:42.360642 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1b31cce6-b000-448f-853e-231f707a2be5" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.213:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 16:46:42 crc kubenswrapper[4814]: I0227 16:46:42.360699 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1b31cce6-b000-448f-853e-231f707a2be5" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.213:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 16:46:44 crc kubenswrapper[4814]: I0227 16:46:44.342229 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 27 16:46:44 crc kubenswrapper[4814]: I0227 16:46:44.342347 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 27 16:46:45 crc kubenswrapper[4814]: I0227 16:46:45.358496 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b7902f26-5f7d-4c34-8166-d1857a631858" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 16:46:45 crc kubenswrapper[4814]: I0227 16:46:45.358520 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b7902f26-5f7d-4c34-8166-d1857a631858" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 16:46:46 crc kubenswrapper[4814]: I0227 16:46:46.120050 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 27 16:46:46 crc kubenswrapper[4814]: I0227 16:46:46.151673 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 27 16:46:46 crc kubenswrapper[4814]: I0227 16:46:46.900816 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 27 16:46:48 crc kubenswrapper[4814]: I0227 16:46:48.134607 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 27 16:46:51 crc kubenswrapper[4814]: I0227 16:46:51.349932 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 27 16:46:51 crc kubenswrapper[4814]: I0227 16:46:51.350593 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 27 16:46:51 crc kubenswrapper[4814]: I0227 16:46:51.350908 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 27 16:46:51 crc kubenswrapper[4814]: I0227 16:46:51.350949 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 27 16:46:51 crc kubenswrapper[4814]: I0227 16:46:51.367947 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 27 16:46:51 crc kubenswrapper[4814]: I0227 16:46:51.369356 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 27 16:46:52 crc kubenswrapper[4814]: I0227 16:46:52.805048 4814 scope.go:117] "RemoveContainer" containerID="3bba3d181a7796362e15415c9ce922f887f8662bee51b24906b208723a196e50" Feb 27 16:46:54 crc kubenswrapper[4814]: I0227 16:46:54.350454 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 27 16:46:54 crc kubenswrapper[4814]: I0227 16:46:54.351210 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 27 16:46:54 crc kubenswrapper[4814]: I0227 16:46:54.361498 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 27 16:46:54 crc kubenswrapper[4814]: I0227 16:46:54.956205 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 27 16:47:03 crc kubenswrapper[4814]: I0227 16:47:03.556338 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 16:47:04 crc kubenswrapper[4814]: I0227 16:47:04.346802 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 16:47:08 crc kubenswrapper[4814]: I0227 16:47:08.123623 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="58660097-6b23-4016-98b4-6a10978c0887" containerName="rabbitmq" containerID="cri-o://a535cf2aaf66daf9350afc5bbf95a9a881e393aac3022afed976df8d33d602f2" gracePeriod=604796 Feb 27 16:47:09 crc kubenswrapper[4814]: I0227 16:47:09.079460 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="8bfc5fb5-c9ab-4108-915b-e6669b332e05" containerName="rabbitmq" containerID="cri-o://0375689c55064914df1dda469516539c22548c73611fd0971ce7f5ec9e6fe3a0" gracePeriod=604796 Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.754686 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.919599 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/58660097-6b23-4016-98b4-6a10978c0887-pod-info\") pod \"58660097-6b23-4016-98b4-6a10978c0887\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.919708 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-erlang-cookie\") pod \"58660097-6b23-4016-98b4-6a10978c0887\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.919734 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-plugins\") pod \"58660097-6b23-4016-98b4-6a10978c0887\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.919768 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-confd\") pod \"58660097-6b23-4016-98b4-6a10978c0887\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.919816 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-plugins-conf\") pod \"58660097-6b23-4016-98b4-6a10978c0887\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.919843 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/58660097-6b23-4016-98b4-6a10978c0887-erlang-cookie-secret\") pod \"58660097-6b23-4016-98b4-6a10978c0887\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.919882 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j42jm\" (UniqueName: \"kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-kube-api-access-j42jm\") pod \"58660097-6b23-4016-98b4-6a10978c0887\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.919930 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-server-conf\") pod \"58660097-6b23-4016-98b4-6a10978c0887\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.920170 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-config-data\") pod \"58660097-6b23-4016-98b4-6a10978c0887\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.920281 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"58660097-6b23-4016-98b4-6a10978c0887\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.920393 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-tls\") pod \"58660097-6b23-4016-98b4-6a10978c0887\" (UID: \"58660097-6b23-4016-98b4-6a10978c0887\") " Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.920814 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "58660097-6b23-4016-98b4-6a10978c0887" (UID: "58660097-6b23-4016-98b4-6a10978c0887"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.921564 4814 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.921685 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "58660097-6b23-4016-98b4-6a10978c0887" (UID: "58660097-6b23-4016-98b4-6a10978c0887"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.922100 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "58660097-6b23-4016-98b4-6a10978c0887" (UID: "58660097-6b23-4016-98b4-6a10978c0887"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.930090 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-kube-api-access-j42jm" (OuterVolumeSpecName: "kube-api-access-j42jm") pod "58660097-6b23-4016-98b4-6a10978c0887" (UID: "58660097-6b23-4016-98b4-6a10978c0887"). InnerVolumeSpecName "kube-api-access-j42jm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.930850 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "58660097-6b23-4016-98b4-6a10978c0887" (UID: "58660097-6b23-4016-98b4-6a10978c0887"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.930863 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "58660097-6b23-4016-98b4-6a10978c0887" (UID: "58660097-6b23-4016-98b4-6a10978c0887"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.930936 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/58660097-6b23-4016-98b4-6a10978c0887-pod-info" (OuterVolumeSpecName: "pod-info") pod "58660097-6b23-4016-98b4-6a10978c0887" (UID: "58660097-6b23-4016-98b4-6a10978c0887"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.931232 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58660097-6b23-4016-98b4-6a10978c0887-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "58660097-6b23-4016-98b4-6a10978c0887" (UID: "58660097-6b23-4016-98b4-6a10978c0887"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.958858 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-config-data" (OuterVolumeSpecName: "config-data") pod "58660097-6b23-4016-98b4-6a10978c0887" (UID: "58660097-6b23-4016-98b4-6a10978c0887"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:14 crc kubenswrapper[4814]: I0227 16:47:14.987174 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-server-conf" (OuterVolumeSpecName: "server-conf") pod "58660097-6b23-4016-98b4-6a10978c0887" (UID: "58660097-6b23-4016-98b4-6a10978c0887"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.024106 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.024169 4814 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.024185 4814 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.024198 4814 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/58660097-6b23-4016-98b4-6a10978c0887-pod-info\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.024215 4814 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.024226 4814 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.024236 4814 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/58660097-6b23-4016-98b4-6a10978c0887-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.024267 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j42jm\" (UniqueName: \"kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-kube-api-access-j42jm\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.024279 4814 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/58660097-6b23-4016-98b4-6a10978c0887-server-conf\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.061811 4814 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.075118 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "58660097-6b23-4016-98b4-6a10978c0887" (UID: "58660097-6b23-4016-98b4-6a10978c0887"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.126146 4814 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.126195 4814 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/58660097-6b23-4016-98b4-6a10978c0887-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.189851 4814 generic.go:334] "Generic (PLEG): container finished" podID="58660097-6b23-4016-98b4-6a10978c0887" containerID="a535cf2aaf66daf9350afc5bbf95a9a881e393aac3022afed976df8d33d602f2" exitCode=0 Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.189898 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"58660097-6b23-4016-98b4-6a10978c0887","Type":"ContainerDied","Data":"a535cf2aaf66daf9350afc5bbf95a9a881e393aac3022afed976df8d33d602f2"} Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.189933 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"58660097-6b23-4016-98b4-6a10978c0887","Type":"ContainerDied","Data":"d12bf0f65745d1088e3f5b639958b7edfd15f8e0f3f1a158a85811b8e0a9c149"} Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.189951 4814 scope.go:117] "RemoveContainer" containerID="a535cf2aaf66daf9350afc5bbf95a9a881e393aac3022afed976df8d33d602f2" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.189963 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.219287 4814 scope.go:117] "RemoveContainer" containerID="3901d5b87aa0457909d2bb6c3ba5880bef5442d1430009b012abf06b49f82fba" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.295544 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.296538 4814 scope.go:117] "RemoveContainer" containerID="a535cf2aaf66daf9350afc5bbf95a9a881e393aac3022afed976df8d33d602f2" Feb 27 16:47:15 crc kubenswrapper[4814]: E0227 16:47:15.297045 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a535cf2aaf66daf9350afc5bbf95a9a881e393aac3022afed976df8d33d602f2\": container with ID starting with a535cf2aaf66daf9350afc5bbf95a9a881e393aac3022afed976df8d33d602f2 not found: ID does not exist" containerID="a535cf2aaf66daf9350afc5bbf95a9a881e393aac3022afed976df8d33d602f2" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.297102 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a535cf2aaf66daf9350afc5bbf95a9a881e393aac3022afed976df8d33d602f2"} err="failed to get container status \"a535cf2aaf66daf9350afc5bbf95a9a881e393aac3022afed976df8d33d602f2\": rpc error: code = NotFound desc = could not find container \"a535cf2aaf66daf9350afc5bbf95a9a881e393aac3022afed976df8d33d602f2\": container with ID starting with a535cf2aaf66daf9350afc5bbf95a9a881e393aac3022afed976df8d33d602f2 not found: ID does not exist" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.297145 4814 scope.go:117] "RemoveContainer" containerID="3901d5b87aa0457909d2bb6c3ba5880bef5442d1430009b012abf06b49f82fba" Feb 27 16:47:15 crc kubenswrapper[4814]: E0227 16:47:15.297991 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3901d5b87aa0457909d2bb6c3ba5880bef5442d1430009b012abf06b49f82fba\": container with ID starting with 3901d5b87aa0457909d2bb6c3ba5880bef5442d1430009b012abf06b49f82fba not found: ID does not exist" containerID="3901d5b87aa0457909d2bb6c3ba5880bef5442d1430009b012abf06b49f82fba" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.298036 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3901d5b87aa0457909d2bb6c3ba5880bef5442d1430009b012abf06b49f82fba"} err="failed to get container status \"3901d5b87aa0457909d2bb6c3ba5880bef5442d1430009b012abf06b49f82fba\": rpc error: code = NotFound desc = could not find container \"3901d5b87aa0457909d2bb6c3ba5880bef5442d1430009b012abf06b49f82fba\": container with ID starting with 3901d5b87aa0457909d2bb6c3ba5880bef5442d1430009b012abf06b49f82fba not found: ID does not exist" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.309144 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.345018 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 16:47:15 crc kubenswrapper[4814]: E0227 16:47:15.345536 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58660097-6b23-4016-98b4-6a10978c0887" containerName="rabbitmq" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.345563 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="58660097-6b23-4016-98b4-6a10978c0887" containerName="rabbitmq" Feb 27 16:47:15 crc kubenswrapper[4814]: E0227 16:47:15.345598 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58660097-6b23-4016-98b4-6a10978c0887" containerName="setup-container" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.345605 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="58660097-6b23-4016-98b4-6a10978c0887" containerName="setup-container" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.345814 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="58660097-6b23-4016-98b4-6a10978c0887" containerName="rabbitmq" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.346808 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.351666 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.351926 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.352073 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.352363 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.352581 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-dg2h7" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.352787 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.374170 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.374282 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.538446 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2fb1eae8-a753-43f3-8ed7-759e6e211ace-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.538744 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.538779 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2fb1eae8-a753-43f3-8ed7-759e6e211ace-config-data\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.538812 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2fb1eae8-a753-43f3-8ed7-759e6e211ace-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.538839 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2fb1eae8-a753-43f3-8ed7-759e6e211ace-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.538854 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2fb1eae8-a753-43f3-8ed7-759e6e211ace-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.538873 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2fb1eae8-a753-43f3-8ed7-759e6e211ace-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.538910 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7jrz\" (UniqueName: \"kubernetes.io/projected/2fb1eae8-a753-43f3-8ed7-759e6e211ace-kube-api-access-w7jrz\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.538936 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2fb1eae8-a753-43f3-8ed7-759e6e211ace-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.538968 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2fb1eae8-a753-43f3-8ed7-759e6e211ace-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.539000 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2fb1eae8-a753-43f3-8ed7-759e6e211ace-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.641092 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2fb1eae8-a753-43f3-8ed7-759e6e211ace-config-data\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.641161 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2fb1eae8-a753-43f3-8ed7-759e6e211ace-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.641195 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2fb1eae8-a753-43f3-8ed7-759e6e211ace-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.641212 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2fb1eae8-a753-43f3-8ed7-759e6e211ace-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.641236 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2fb1eae8-a753-43f3-8ed7-759e6e211ace-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.641299 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7jrz\" (UniqueName: \"kubernetes.io/projected/2fb1eae8-a753-43f3-8ed7-759e6e211ace-kube-api-access-w7jrz\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.641337 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2fb1eae8-a753-43f3-8ed7-759e6e211ace-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.641381 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2fb1eae8-a753-43f3-8ed7-759e6e211ace-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.641439 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2fb1eae8-a753-43f3-8ed7-759e6e211ace-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.641469 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2fb1eae8-a753-43f3-8ed7-759e6e211ace-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.641525 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.642105 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.642117 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2fb1eae8-a753-43f3-8ed7-759e6e211ace-config-data\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.642485 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2fb1eae8-a753-43f3-8ed7-759e6e211ace-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.642597 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2fb1eae8-a753-43f3-8ed7-759e6e211ace-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.642743 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2fb1eae8-a753-43f3-8ed7-759e6e211ace-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.643205 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2fb1eae8-a753-43f3-8ed7-759e6e211ace-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.646898 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2fb1eae8-a753-43f3-8ed7-759e6e211ace-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.647761 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2fb1eae8-a753-43f3-8ed7-759e6e211ace-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.648398 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2fb1eae8-a753-43f3-8ed7-759e6e211ace-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.648497 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2fb1eae8-a753-43f3-8ed7-759e6e211ace-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.660592 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7jrz\" (UniqueName: \"kubernetes.io/projected/2fb1eae8-a753-43f3-8ed7-759e6e211ace-kube-api-access-w7jrz\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.673645 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"2fb1eae8-a753-43f3-8ed7-759e6e211ace\") " pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.746701 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.752141 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.844952 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-tls\") pod \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.845109 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-plugins-conf\") pod \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.845148 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ll9d5\" (UniqueName: \"kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-kube-api-access-ll9d5\") pod \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.845225 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-plugins\") pod \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.845321 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-erlang-cookie\") pod \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.845391 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8bfc5fb5-c9ab-4108-915b-e6669b332e05-pod-info\") pod \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.845447 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-config-data\") pod \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.845530 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.845557 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-confd\") pod \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.845585 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8bfc5fb5-c9ab-4108-915b-e6669b332e05-erlang-cookie-secret\") pod \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.845620 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-server-conf\") pod \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\" (UID: \"8bfc5fb5-c9ab-4108-915b-e6669b332e05\") " Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.845777 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "8bfc5fb5-c9ab-4108-915b-e6669b332e05" (UID: "8bfc5fb5-c9ab-4108-915b-e6669b332e05"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.846059 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "8bfc5fb5-c9ab-4108-915b-e6669b332e05" (UID: "8bfc5fb5-c9ab-4108-915b-e6669b332e05"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.846671 4814 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.846693 4814 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.848345 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "8bfc5fb5-c9ab-4108-915b-e6669b332e05" (UID: "8bfc5fb5-c9ab-4108-915b-e6669b332e05"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.848808 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "8bfc5fb5-c9ab-4108-915b-e6669b332e05" (UID: "8bfc5fb5-c9ab-4108-915b-e6669b332e05"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.850100 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-kube-api-access-ll9d5" (OuterVolumeSpecName: "kube-api-access-ll9d5") pod "8bfc5fb5-c9ab-4108-915b-e6669b332e05" (UID: "8bfc5fb5-c9ab-4108-915b-e6669b332e05"). InnerVolumeSpecName "kube-api-access-ll9d5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.851345 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/8bfc5fb5-c9ab-4108-915b-e6669b332e05-pod-info" (OuterVolumeSpecName: "pod-info") pod "8bfc5fb5-c9ab-4108-915b-e6669b332e05" (UID: "8bfc5fb5-c9ab-4108-915b-e6669b332e05"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.855881 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "8bfc5fb5-c9ab-4108-915b-e6669b332e05" (UID: "8bfc5fb5-c9ab-4108-915b-e6669b332e05"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.857598 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bfc5fb5-c9ab-4108-915b-e6669b332e05-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "8bfc5fb5-c9ab-4108-915b-e6669b332e05" (UID: "8bfc5fb5-c9ab-4108-915b-e6669b332e05"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.881208 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-config-data" (OuterVolumeSpecName: "config-data") pod "8bfc5fb5-c9ab-4108-915b-e6669b332e05" (UID: "8bfc5fb5-c9ab-4108-915b-e6669b332e05"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.917532 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-server-conf" (OuterVolumeSpecName: "server-conf") pod "8bfc5fb5-c9ab-4108-915b-e6669b332e05" (UID: "8bfc5fb5-c9ab-4108-915b-e6669b332e05"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.953242 4814 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8bfc5fb5-c9ab-4108-915b-e6669b332e05-pod-info\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.953287 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.953312 4814 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.953321 4814 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8bfc5fb5-c9ab-4108-915b-e6669b332e05-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.953332 4814 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8bfc5fb5-c9ab-4108-915b-e6669b332e05-server-conf\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.953340 4814 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.953350 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ll9d5\" (UniqueName: \"kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-kube-api-access-ll9d5\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.953361 4814 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:15 crc kubenswrapper[4814]: I0227 16:47:15.985663 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "8bfc5fb5-c9ab-4108-915b-e6669b332e05" (UID: "8bfc5fb5-c9ab-4108-915b-e6669b332e05"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.004877 4814 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.006110 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-5vgcq"] Feb 27 16:47:16 crc kubenswrapper[4814]: E0227 16:47:16.006645 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bfc5fb5-c9ab-4108-915b-e6669b332e05" containerName="setup-container" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.006669 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bfc5fb5-c9ab-4108-915b-e6669b332e05" containerName="setup-container" Feb 27 16:47:16 crc kubenswrapper[4814]: E0227 16:47:16.006686 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bfc5fb5-c9ab-4108-915b-e6669b332e05" containerName="rabbitmq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.006693 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bfc5fb5-c9ab-4108-915b-e6669b332e05" containerName="rabbitmq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.006904 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bfc5fb5-c9ab-4108-915b-e6669b332e05" containerName="rabbitmq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.007992 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.013542 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.025432 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-5vgcq"] Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.048575 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.056047 4814 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.056075 4814 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8bfc5fb5-c9ab-4108-915b-e6669b332e05-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.157759 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4wsv\" (UniqueName: \"kubernetes.io/projected/ae9a0fcd-a023-4879-894e-55a80628bb6c-kube-api-access-w4wsv\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.157804 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-config\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.157848 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.157877 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.158105 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.158374 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.158529 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-dns-svc\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.199212 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2fb1eae8-a753-43f3-8ed7-759e6e211ace","Type":"ContainerStarted","Data":"d45929edf1da38f98b6fcbd32ee12a0e59987dfcb4e546a3bab3c61738389e52"} Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.201223 4814 generic.go:334] "Generic (PLEG): container finished" podID="8bfc5fb5-c9ab-4108-915b-e6669b332e05" containerID="0375689c55064914df1dda469516539c22548c73611fd0971ce7f5ec9e6fe3a0" exitCode=0 Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.201265 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8bfc5fb5-c9ab-4108-915b-e6669b332e05","Type":"ContainerDied","Data":"0375689c55064914df1dda469516539c22548c73611fd0971ce7f5ec9e6fe3a0"} Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.201282 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8bfc5fb5-c9ab-4108-915b-e6669b332e05","Type":"ContainerDied","Data":"fd5105f23466970276cfc746c6b0fd5d2ec801f03303da64073e4d930913f7eb"} Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.201298 4814 scope.go:117] "RemoveContainer" containerID="0375689c55064914df1dda469516539c22548c73611fd0971ce7f5ec9e6fe3a0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.201408 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.233571 4814 scope.go:117] "RemoveContainer" containerID="2c129c7f4218bd10e27a0ca4ecd6c565a206b1a457efa3f4a5f5c065833f539d" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.233694 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.247973 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.266474 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-dns-svc\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.266552 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4wsv\" (UniqueName: \"kubernetes.io/projected/ae9a0fcd-a023-4879-894e-55a80628bb6c-kube-api-access-w4wsv\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.266576 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-config\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.266602 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.266625 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.266672 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.266720 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.267466 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.267970 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-dns-svc\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.268754 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-config\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.269290 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.269980 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.270069 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.274667 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.276803 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.279867 4814 scope.go:117] "RemoveContainer" containerID="0375689c55064914df1dda469516539c22548c73611fd0971ce7f5ec9e6fe3a0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.280980 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.281559 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.281747 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.282716 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-8xdnh" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.282884 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.283167 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.283389 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 27 16:47:16 crc kubenswrapper[4814]: E0227 16:47:16.284149 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0375689c55064914df1dda469516539c22548c73611fd0971ce7f5ec9e6fe3a0\": container with ID starting with 0375689c55064914df1dda469516539c22548c73611fd0971ce7f5ec9e6fe3a0 not found: ID does not exist" containerID="0375689c55064914df1dda469516539c22548c73611fd0971ce7f5ec9e6fe3a0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.284197 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0375689c55064914df1dda469516539c22548c73611fd0971ce7f5ec9e6fe3a0"} err="failed to get container status \"0375689c55064914df1dda469516539c22548c73611fd0971ce7f5ec9e6fe3a0\": rpc error: code = NotFound desc = could not find container \"0375689c55064914df1dda469516539c22548c73611fd0971ce7f5ec9e6fe3a0\": container with ID starting with 0375689c55064914df1dda469516539c22548c73611fd0971ce7f5ec9e6fe3a0 not found: ID does not exist" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.284228 4814 scope.go:117] "RemoveContainer" containerID="2c129c7f4218bd10e27a0ca4ecd6c565a206b1a457efa3f4a5f5c065833f539d" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.284425 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 16:47:16 crc kubenswrapper[4814]: E0227 16:47:16.284671 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c129c7f4218bd10e27a0ca4ecd6c565a206b1a457efa3f4a5f5c065833f539d\": container with ID starting with 2c129c7f4218bd10e27a0ca4ecd6c565a206b1a457efa3f4a5f5c065833f539d not found: ID does not exist" containerID="2c129c7f4218bd10e27a0ca4ecd6c565a206b1a457efa3f4a5f5c065833f539d" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.284702 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c129c7f4218bd10e27a0ca4ecd6c565a206b1a457efa3f4a5f5c065833f539d"} err="failed to get container status \"2c129c7f4218bd10e27a0ca4ecd6c565a206b1a457efa3f4a5f5c065833f539d\": rpc error: code = NotFound desc = could not find container \"2c129c7f4218bd10e27a0ca4ecd6c565a206b1a457efa3f4a5f5c065833f539d\": container with ID starting with 2c129c7f4218bd10e27a0ca4ecd6c565a206b1a457efa3f4a5f5c065833f539d not found: ID does not exist" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.297455 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4wsv\" (UniqueName: \"kubernetes.io/projected/ae9a0fcd-a023-4879-894e-55a80628bb6c-kube-api-access-w4wsv\") pod \"dnsmasq-dns-67b789f86c-5vgcq\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.341348 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.470185 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/51ca92ea-753f-4e44-94a8-68e73d165193-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.470227 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/51ca92ea-753f-4e44-94a8-68e73d165193-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.470249 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/51ca92ea-753f-4e44-94a8-68e73d165193-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.470276 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/51ca92ea-753f-4e44-94a8-68e73d165193-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.470294 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/51ca92ea-753f-4e44-94a8-68e73d165193-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.470493 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgf8h\" (UniqueName: \"kubernetes.io/projected/51ca92ea-753f-4e44-94a8-68e73d165193-kube-api-access-dgf8h\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.470714 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/51ca92ea-753f-4e44-94a8-68e73d165193-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.470740 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/51ca92ea-753f-4e44-94a8-68e73d165193-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.470822 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/51ca92ea-753f-4e44-94a8-68e73d165193-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.470853 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/51ca92ea-753f-4e44-94a8-68e73d165193-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.470897 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.504312 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58660097-6b23-4016-98b4-6a10978c0887" path="/var/lib/kubelet/pods/58660097-6b23-4016-98b4-6a10978c0887/volumes" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.505543 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bfc5fb5-c9ab-4108-915b-e6669b332e05" path="/var/lib/kubelet/pods/8bfc5fb5-c9ab-4108-915b-e6669b332e05/volumes" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.572448 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/51ca92ea-753f-4e44-94a8-68e73d165193-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.572487 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/51ca92ea-753f-4e44-94a8-68e73d165193-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.572522 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/51ca92ea-753f-4e44-94a8-68e73d165193-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.572541 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/51ca92ea-753f-4e44-94a8-68e73d165193-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.572569 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.572625 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/51ca92ea-753f-4e44-94a8-68e73d165193-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.572647 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/51ca92ea-753f-4e44-94a8-68e73d165193-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.572681 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/51ca92ea-753f-4e44-94a8-68e73d165193-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.572714 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/51ca92ea-753f-4e44-94a8-68e73d165193-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.572734 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/51ca92ea-753f-4e44-94a8-68e73d165193-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.572761 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgf8h\" (UniqueName: \"kubernetes.io/projected/51ca92ea-753f-4e44-94a8-68e73d165193-kube-api-access-dgf8h\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.573323 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/51ca92ea-753f-4e44-94a8-68e73d165193-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.573446 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/51ca92ea-753f-4e44-94a8-68e73d165193-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.573527 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/51ca92ea-753f-4e44-94a8-68e73d165193-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.573549 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.574021 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/51ca92ea-753f-4e44-94a8-68e73d165193-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.575176 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/51ca92ea-753f-4e44-94a8-68e73d165193-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.577618 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/51ca92ea-753f-4e44-94a8-68e73d165193-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.578542 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/51ca92ea-753f-4e44-94a8-68e73d165193-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.578685 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/51ca92ea-753f-4e44-94a8-68e73d165193-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.583369 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/51ca92ea-753f-4e44-94a8-68e73d165193-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.599433 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgf8h\" (UniqueName: \"kubernetes.io/projected/51ca92ea-753f-4e44-94a8-68e73d165193-kube-api-access-dgf8h\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.608170 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ca92ea-753f-4e44-94a8-68e73d165193\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.811007 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-5vgcq"] Feb 27 16:47:16 crc kubenswrapper[4814]: W0227 16:47:16.811495 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae9a0fcd_a023_4879_894e_55a80628bb6c.slice/crio-381dd244ebf1932472d507d932b82df49f5b532ca5d6f146804f161d15abceb2 WatchSource:0}: Error finding container 381dd244ebf1932472d507d932b82df49f5b532ca5d6f146804f161d15abceb2: Status 404 returned error can't find the container with id 381dd244ebf1932472d507d932b82df49f5b532ca5d6f146804f161d15abceb2 Feb 27 16:47:16 crc kubenswrapper[4814]: I0227 16:47:16.900206 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:17 crc kubenswrapper[4814]: I0227 16:47:17.215488 4814 generic.go:334] "Generic (PLEG): container finished" podID="ae9a0fcd-a023-4879-894e-55a80628bb6c" containerID="d702e165e39ff759e65f7b8841a967f47a414c2beb675786cc2f24a27dcdafc0" exitCode=0 Feb 27 16:47:17 crc kubenswrapper[4814]: I0227 16:47:17.215536 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" event={"ID":"ae9a0fcd-a023-4879-894e-55a80628bb6c","Type":"ContainerDied","Data":"d702e165e39ff759e65f7b8841a967f47a414c2beb675786cc2f24a27dcdafc0"} Feb 27 16:47:17 crc kubenswrapper[4814]: I0227 16:47:17.215772 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" event={"ID":"ae9a0fcd-a023-4879-894e-55a80628bb6c","Type":"ContainerStarted","Data":"381dd244ebf1932472d507d932b82df49f5b532ca5d6f146804f161d15abceb2"} Feb 27 16:47:17 crc kubenswrapper[4814]: I0227 16:47:17.471951 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 16:47:18 crc kubenswrapper[4814]: I0227 16:47:18.233752 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"51ca92ea-753f-4e44-94a8-68e73d165193","Type":"ContainerStarted","Data":"1c5c8225d32e0cd4c24b19056549226202e4954601b7d1268b039a9efb1ee0ec"} Feb 27 16:47:18 crc kubenswrapper[4814]: I0227 16:47:18.236006 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2fb1eae8-a753-43f3-8ed7-759e6e211ace","Type":"ContainerStarted","Data":"19b207db2f45f24913dcbee1a5e62d57c788d325c70cbf5c4b41209b5dab6908"} Feb 27 16:47:18 crc kubenswrapper[4814]: I0227 16:47:18.241762 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" event={"ID":"ae9a0fcd-a023-4879-894e-55a80628bb6c","Type":"ContainerStarted","Data":"905ebb566f56f8c38b241e86291cbeaa889e1cb574d74a93ad3a84f357156388"} Feb 27 16:47:18 crc kubenswrapper[4814]: I0227 16:47:18.242182 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:18 crc kubenswrapper[4814]: I0227 16:47:18.293945 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" podStartSLOduration=3.293926908 podStartE2EDuration="3.293926908s" podCreationTimestamp="2026-02-27 16:47:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:47:18.29302893 +0000 UTC m=+1450.745653790" watchObservedRunningTime="2026-02-27 16:47:18.293926908 +0000 UTC m=+1450.746551748" Feb 27 16:47:20 crc kubenswrapper[4814]: I0227 16:47:20.270113 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"51ca92ea-753f-4e44-94a8-68e73d165193","Type":"ContainerStarted","Data":"3c171ce5597e0c2dc06c8e8eddb953dea1ee23d2a0bc2932815d090eef1b7be0"} Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.343608 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.474876 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-vbvs4"] Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.475108 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" podUID="2bb45fed-87b8-44f9-8afb-659219445949" containerName="dnsmasq-dns" containerID="cri-o://15849b9d9c0da80aec16722483b871cd48f80d72e782fd62e42d89a0c01a4065" gracePeriod=10 Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.561275 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-tww6q"] Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.562799 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.585582 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-tww6q"] Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.719732 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.719803 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.719825 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.719877 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-config\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.719966 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlm74\" (UniqueName: \"kubernetes.io/projected/18e28b72-72be-49fd-b568-c3a396e87db0-kube-api-access-wlm74\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.719990 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.720028 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.821598 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-config\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.821998 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlm74\" (UniqueName: \"kubernetes.io/projected/18e28b72-72be-49fd-b568-c3a396e87db0-kube-api-access-wlm74\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.822024 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.822082 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.822124 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.822191 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.822210 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.822594 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-config\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.823236 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.823371 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.823446 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.823484 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.824118 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18e28b72-72be-49fd-b568-c3a396e87db0-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.844474 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlm74\" (UniqueName: \"kubernetes.io/projected/18e28b72-72be-49fd-b568-c3a396e87db0-kube-api-access-wlm74\") pod \"dnsmasq-dns-cb6ffcf87-tww6q\" (UID: \"18e28b72-72be-49fd-b568-c3a396e87db0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.884848 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:26 crc kubenswrapper[4814]: I0227 16:47:26.978816 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.126653 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-dns-svc\") pod \"2bb45fed-87b8-44f9-8afb-659219445949\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.126735 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pggk\" (UniqueName: \"kubernetes.io/projected/2bb45fed-87b8-44f9-8afb-659219445949-kube-api-access-5pggk\") pod \"2bb45fed-87b8-44f9-8afb-659219445949\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.126794 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-ovsdbserver-sb\") pod \"2bb45fed-87b8-44f9-8afb-659219445949\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.126836 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-config\") pod \"2bb45fed-87b8-44f9-8afb-659219445949\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.126883 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-dns-swift-storage-0\") pod \"2bb45fed-87b8-44f9-8afb-659219445949\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.126941 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-ovsdbserver-nb\") pod \"2bb45fed-87b8-44f9-8afb-659219445949\" (UID: \"2bb45fed-87b8-44f9-8afb-659219445949\") " Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.131488 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bb45fed-87b8-44f9-8afb-659219445949-kube-api-access-5pggk" (OuterVolumeSpecName: "kube-api-access-5pggk") pod "2bb45fed-87b8-44f9-8afb-659219445949" (UID: "2bb45fed-87b8-44f9-8afb-659219445949"). InnerVolumeSpecName "kube-api-access-5pggk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.189650 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2bb45fed-87b8-44f9-8afb-659219445949" (UID: "2bb45fed-87b8-44f9-8afb-659219445949"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.190378 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2bb45fed-87b8-44f9-8afb-659219445949" (UID: "2bb45fed-87b8-44f9-8afb-659219445949"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.202177 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2bb45fed-87b8-44f9-8afb-659219445949" (UID: "2bb45fed-87b8-44f9-8afb-659219445949"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.203489 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2bb45fed-87b8-44f9-8afb-659219445949" (UID: "2bb45fed-87b8-44f9-8afb-659219445949"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.218079 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-config" (OuterVolumeSpecName: "config") pod "2bb45fed-87b8-44f9-8afb-659219445949" (UID: "2bb45fed-87b8-44f9-8afb-659219445949"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.229025 4814 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.229054 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pggk\" (UniqueName: \"kubernetes.io/projected/2bb45fed-87b8-44f9-8afb-659219445949-kube-api-access-5pggk\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.229066 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.229075 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.229084 4814 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.229091 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bb45fed-87b8-44f9-8afb-659219445949-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.367048 4814 generic.go:334] "Generic (PLEG): container finished" podID="2bb45fed-87b8-44f9-8afb-659219445949" containerID="15849b9d9c0da80aec16722483b871cd48f80d72e782fd62e42d89a0c01a4065" exitCode=0 Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.367091 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" event={"ID":"2bb45fed-87b8-44f9-8afb-659219445949","Type":"ContainerDied","Data":"15849b9d9c0da80aec16722483b871cd48f80d72e782fd62e42d89a0c01a4065"} Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.367125 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" event={"ID":"2bb45fed-87b8-44f9-8afb-659219445949","Type":"ContainerDied","Data":"b1f70f155b1c700d6f1373ab25d36b9e9e9f0c9ac0c5746628ef194365ec25ea"} Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.367143 4814 scope.go:117] "RemoveContainer" containerID="15849b9d9c0da80aec16722483b871cd48f80d72e782fd62e42d89a0c01a4065" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.367151 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-vbvs4" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.398948 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-tww6q"] Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.430594 4814 scope.go:117] "RemoveContainer" containerID="10cbc2745d2e87aa081224dadf4b97d4c4d63f41466a76d95c1d6d9b48dffd83" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.432794 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-vbvs4"] Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.441926 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-vbvs4"] Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.465641 4814 scope.go:117] "RemoveContainer" containerID="15849b9d9c0da80aec16722483b871cd48f80d72e782fd62e42d89a0c01a4065" Feb 27 16:47:27 crc kubenswrapper[4814]: E0227 16:47:27.466409 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15849b9d9c0da80aec16722483b871cd48f80d72e782fd62e42d89a0c01a4065\": container with ID starting with 15849b9d9c0da80aec16722483b871cd48f80d72e782fd62e42d89a0c01a4065 not found: ID does not exist" containerID="15849b9d9c0da80aec16722483b871cd48f80d72e782fd62e42d89a0c01a4065" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.466448 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15849b9d9c0da80aec16722483b871cd48f80d72e782fd62e42d89a0c01a4065"} err="failed to get container status \"15849b9d9c0da80aec16722483b871cd48f80d72e782fd62e42d89a0c01a4065\": rpc error: code = NotFound desc = could not find container \"15849b9d9c0da80aec16722483b871cd48f80d72e782fd62e42d89a0c01a4065\": container with ID starting with 15849b9d9c0da80aec16722483b871cd48f80d72e782fd62e42d89a0c01a4065 not found: ID does not exist" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.466476 4814 scope.go:117] "RemoveContainer" containerID="10cbc2745d2e87aa081224dadf4b97d4c4d63f41466a76d95c1d6d9b48dffd83" Feb 27 16:47:27 crc kubenswrapper[4814]: E0227 16:47:27.467685 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10cbc2745d2e87aa081224dadf4b97d4c4d63f41466a76d95c1d6d9b48dffd83\": container with ID starting with 10cbc2745d2e87aa081224dadf4b97d4c4d63f41466a76d95c1d6d9b48dffd83 not found: ID does not exist" containerID="10cbc2745d2e87aa081224dadf4b97d4c4d63f41466a76d95c1d6d9b48dffd83" Feb 27 16:47:27 crc kubenswrapper[4814]: I0227 16:47:27.467738 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10cbc2745d2e87aa081224dadf4b97d4c4d63f41466a76d95c1d6d9b48dffd83"} err="failed to get container status \"10cbc2745d2e87aa081224dadf4b97d4c4d63f41466a76d95c1d6d9b48dffd83\": rpc error: code = NotFound desc = could not find container \"10cbc2745d2e87aa081224dadf4b97d4c4d63f41466a76d95c1d6d9b48dffd83\": container with ID starting with 10cbc2745d2e87aa081224dadf4b97d4c4d63f41466a76d95c1d6d9b48dffd83 not found: ID does not exist" Feb 27 16:47:28 crc kubenswrapper[4814]: I0227 16:47:28.386940 4814 generic.go:334] "Generic (PLEG): container finished" podID="18e28b72-72be-49fd-b568-c3a396e87db0" containerID="890383e07308bfe65e3863895ff82a22f5f743f0d96d66319166389dafc04a2a" exitCode=0 Feb 27 16:47:28 crc kubenswrapper[4814]: I0227 16:47:28.386999 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" event={"ID":"18e28b72-72be-49fd-b568-c3a396e87db0","Type":"ContainerDied","Data":"890383e07308bfe65e3863895ff82a22f5f743f0d96d66319166389dafc04a2a"} Feb 27 16:47:28 crc kubenswrapper[4814]: I0227 16:47:28.387483 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" event={"ID":"18e28b72-72be-49fd-b568-c3a396e87db0","Type":"ContainerStarted","Data":"7d73e60f35536c1ba44559a4445da7b6b00e35faac4e823ddb7ecf962698fc85"} Feb 27 16:47:28 crc kubenswrapper[4814]: I0227 16:47:28.517927 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bb45fed-87b8-44f9-8afb-659219445949" path="/var/lib/kubelet/pods/2bb45fed-87b8-44f9-8afb-659219445949/volumes" Feb 27 16:47:29 crc kubenswrapper[4814]: I0227 16:47:29.412306 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" event={"ID":"18e28b72-72be-49fd-b568-c3a396e87db0","Type":"ContainerStarted","Data":"97acc3309b73effcfb74c6de7bdf331c9b929daa46571c51723645b5a1d732de"} Feb 27 16:47:29 crc kubenswrapper[4814]: I0227 16:47:29.412764 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:29 crc kubenswrapper[4814]: I0227 16:47:29.454078 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" podStartSLOduration=3.454059225 podStartE2EDuration="3.454059225s" podCreationTimestamp="2026-02-27 16:47:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:47:29.443551826 +0000 UTC m=+1461.896176676" watchObservedRunningTime="2026-02-27 16:47:29.454059225 +0000 UTC m=+1461.906684075" Feb 27 16:47:36 crc kubenswrapper[4814]: I0227 16:47:36.887607 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6ffcf87-tww6q" Feb 27 16:47:36 crc kubenswrapper[4814]: I0227 16:47:36.984558 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-5vgcq"] Feb 27 16:47:36 crc kubenswrapper[4814]: I0227 16:47:36.984851 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" podUID="ae9a0fcd-a023-4879-894e-55a80628bb6c" containerName="dnsmasq-dns" containerID="cri-o://905ebb566f56f8c38b241e86291cbeaa889e1cb574d74a93ad3a84f357156388" gracePeriod=10 Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.457746 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.507678 4814 generic.go:334] "Generic (PLEG): container finished" podID="ae9a0fcd-a023-4879-894e-55a80628bb6c" containerID="905ebb566f56f8c38b241e86291cbeaa889e1cb574d74a93ad3a84f357156388" exitCode=0 Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.507725 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" event={"ID":"ae9a0fcd-a023-4879-894e-55a80628bb6c","Type":"ContainerDied","Data":"905ebb566f56f8c38b241e86291cbeaa889e1cb574d74a93ad3a84f357156388"} Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.507751 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" event={"ID":"ae9a0fcd-a023-4879-894e-55a80628bb6c","Type":"ContainerDied","Data":"381dd244ebf1932472d507d932b82df49f5b532ca5d6f146804f161d15abceb2"} Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.507767 4814 scope.go:117] "RemoveContainer" containerID="905ebb566f56f8c38b241e86291cbeaa889e1cb574d74a93ad3a84f357156388" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.507892 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-5vgcq" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.540509 4814 scope.go:117] "RemoveContainer" containerID="d702e165e39ff759e65f7b8841a967f47a414c2beb675786cc2f24a27dcdafc0" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.563623 4814 scope.go:117] "RemoveContainer" containerID="905ebb566f56f8c38b241e86291cbeaa889e1cb574d74a93ad3a84f357156388" Feb 27 16:47:37 crc kubenswrapper[4814]: E0227 16:47:37.564680 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"905ebb566f56f8c38b241e86291cbeaa889e1cb574d74a93ad3a84f357156388\": container with ID starting with 905ebb566f56f8c38b241e86291cbeaa889e1cb574d74a93ad3a84f357156388 not found: ID does not exist" containerID="905ebb566f56f8c38b241e86291cbeaa889e1cb574d74a93ad3a84f357156388" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.564713 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"905ebb566f56f8c38b241e86291cbeaa889e1cb574d74a93ad3a84f357156388"} err="failed to get container status \"905ebb566f56f8c38b241e86291cbeaa889e1cb574d74a93ad3a84f357156388\": rpc error: code = NotFound desc = could not find container \"905ebb566f56f8c38b241e86291cbeaa889e1cb574d74a93ad3a84f357156388\": container with ID starting with 905ebb566f56f8c38b241e86291cbeaa889e1cb574d74a93ad3a84f357156388 not found: ID does not exist" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.564734 4814 scope.go:117] "RemoveContainer" containerID="d702e165e39ff759e65f7b8841a967f47a414c2beb675786cc2f24a27dcdafc0" Feb 27 16:47:37 crc kubenswrapper[4814]: E0227 16:47:37.565089 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d702e165e39ff759e65f7b8841a967f47a414c2beb675786cc2f24a27dcdafc0\": container with ID starting with d702e165e39ff759e65f7b8841a967f47a414c2beb675786cc2f24a27dcdafc0 not found: ID does not exist" containerID="d702e165e39ff759e65f7b8841a967f47a414c2beb675786cc2f24a27dcdafc0" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.565132 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d702e165e39ff759e65f7b8841a967f47a414c2beb675786cc2f24a27dcdafc0"} err="failed to get container status \"d702e165e39ff759e65f7b8841a967f47a414c2beb675786cc2f24a27dcdafc0\": rpc error: code = NotFound desc = could not find container \"d702e165e39ff759e65f7b8841a967f47a414c2beb675786cc2f24a27dcdafc0\": container with ID starting with d702e165e39ff759e65f7b8841a967f47a414c2beb675786cc2f24a27dcdafc0 not found: ID does not exist" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.610985 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4wsv\" (UniqueName: \"kubernetes.io/projected/ae9a0fcd-a023-4879-894e-55a80628bb6c-kube-api-access-w4wsv\") pod \"ae9a0fcd-a023-4879-894e-55a80628bb6c\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.611038 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-dns-swift-storage-0\") pod \"ae9a0fcd-a023-4879-894e-55a80628bb6c\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.611083 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-config\") pod \"ae9a0fcd-a023-4879-894e-55a80628bb6c\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.611105 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-dns-svc\") pod \"ae9a0fcd-a023-4879-894e-55a80628bb6c\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.611126 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-openstack-edpm-ipam\") pod \"ae9a0fcd-a023-4879-894e-55a80628bb6c\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.611179 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-ovsdbserver-sb\") pod \"ae9a0fcd-a023-4879-894e-55a80628bb6c\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.611300 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-ovsdbserver-nb\") pod \"ae9a0fcd-a023-4879-894e-55a80628bb6c\" (UID: \"ae9a0fcd-a023-4879-894e-55a80628bb6c\") " Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.616935 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae9a0fcd-a023-4879-894e-55a80628bb6c-kube-api-access-w4wsv" (OuterVolumeSpecName: "kube-api-access-w4wsv") pod "ae9a0fcd-a023-4879-894e-55a80628bb6c" (UID: "ae9a0fcd-a023-4879-894e-55a80628bb6c"). InnerVolumeSpecName "kube-api-access-w4wsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.657405 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ae9a0fcd-a023-4879-894e-55a80628bb6c" (UID: "ae9a0fcd-a023-4879-894e-55a80628bb6c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.657648 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ae9a0fcd-a023-4879-894e-55a80628bb6c" (UID: "ae9a0fcd-a023-4879-894e-55a80628bb6c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.658666 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ae9a0fcd-a023-4879-894e-55a80628bb6c" (UID: "ae9a0fcd-a023-4879-894e-55a80628bb6c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.671321 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "ae9a0fcd-a023-4879-894e-55a80628bb6c" (UID: "ae9a0fcd-a023-4879-894e-55a80628bb6c"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.677013 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-config" (OuterVolumeSpecName: "config") pod "ae9a0fcd-a023-4879-894e-55a80628bb6c" (UID: "ae9a0fcd-a023-4879-894e-55a80628bb6c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.682068 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ae9a0fcd-a023-4879-894e-55a80628bb6c" (UID: "ae9a0fcd-a023-4879-894e-55a80628bb6c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.713751 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4wsv\" (UniqueName: \"kubernetes.io/projected/ae9a0fcd-a023-4879-894e-55a80628bb6c-kube-api-access-w4wsv\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.713779 4814 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.713788 4814 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-config\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.713798 4814 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.713806 4814 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.713815 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.713825 4814 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae9a0fcd-a023-4879-894e-55a80628bb6c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.851833 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-5vgcq"] Feb 27 16:47:37 crc kubenswrapper[4814]: I0227 16:47:37.860699 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-5vgcq"] Feb 27 16:47:38 crc kubenswrapper[4814]: I0227 16:47:38.501649 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae9a0fcd-a023-4879-894e-55a80628bb6c" path="/var/lib/kubelet/pods/ae9a0fcd-a023-4879-894e-55a80628bb6c/volumes" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.152733 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg"] Feb 27 16:47:50 crc kubenswrapper[4814]: E0227 16:47:50.153914 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae9a0fcd-a023-4879-894e-55a80628bb6c" containerName="init" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.153935 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae9a0fcd-a023-4879-894e-55a80628bb6c" containerName="init" Feb 27 16:47:50 crc kubenswrapper[4814]: E0227 16:47:50.153974 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae9a0fcd-a023-4879-894e-55a80628bb6c" containerName="dnsmasq-dns" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.153987 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae9a0fcd-a023-4879-894e-55a80628bb6c" containerName="dnsmasq-dns" Feb 27 16:47:50 crc kubenswrapper[4814]: E0227 16:47:50.154036 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bb45fed-87b8-44f9-8afb-659219445949" containerName="dnsmasq-dns" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.154055 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bb45fed-87b8-44f9-8afb-659219445949" containerName="dnsmasq-dns" Feb 27 16:47:50 crc kubenswrapper[4814]: E0227 16:47:50.154079 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bb45fed-87b8-44f9-8afb-659219445949" containerName="init" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.154090 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bb45fed-87b8-44f9-8afb-659219445949" containerName="init" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.154470 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bb45fed-87b8-44f9-8afb-659219445949" containerName="dnsmasq-dns" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.154650 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae9a0fcd-a023-4879-894e-55a80628bb6c" containerName="dnsmasq-dns" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.155706 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.160876 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.161503 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.161656 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.167631 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg"] Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.169149 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.290955 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg\" (UID: \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.291064 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vczvq\" (UniqueName: \"kubernetes.io/projected/2bce282e-cd29-40eb-bb24-4ba8503d8a38-kube-api-access-vczvq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg\" (UID: \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.291120 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg\" (UID: \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.291148 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg\" (UID: \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.393264 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg\" (UID: \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.393385 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg\" (UID: \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.393520 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vczvq\" (UniqueName: \"kubernetes.io/projected/2bce282e-cd29-40eb-bb24-4ba8503d8a38-kube-api-access-vczvq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg\" (UID: \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.393933 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg\" (UID: \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.399602 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg\" (UID: \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.400324 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg\" (UID: \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.407051 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg\" (UID: \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.426665 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vczvq\" (UniqueName: \"kubernetes.io/projected/2bce282e-cd29-40eb-bb24-4ba8503d8a38-kube-api-access-vczvq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg\" (UID: \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.475649 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.693670 4814 generic.go:334] "Generic (PLEG): container finished" podID="2fb1eae8-a753-43f3-8ed7-759e6e211ace" containerID="19b207db2f45f24913dcbee1a5e62d57c788d325c70cbf5c4b41209b5dab6908" exitCode=0 Feb 27 16:47:50 crc kubenswrapper[4814]: I0227 16:47:50.693714 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2fb1eae8-a753-43f3-8ed7-759e6e211ace","Type":"ContainerDied","Data":"19b207db2f45f24913dcbee1a5e62d57c788d325c70cbf5c4b41209b5dab6908"} Feb 27 16:47:51 crc kubenswrapper[4814]: I0227 16:47:51.079402 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg"] Feb 27 16:47:51 crc kubenswrapper[4814]: W0227 16:47:51.083376 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bce282e_cd29_40eb_bb24_4ba8503d8a38.slice/crio-854534f3125369c349d87e7fb79a36b84a469a1209e020c8fcbe53e8f902b78c WatchSource:0}: Error finding container 854534f3125369c349d87e7fb79a36b84a469a1209e020c8fcbe53e8f902b78c: Status 404 returned error can't find the container with id 854534f3125369c349d87e7fb79a36b84a469a1209e020c8fcbe53e8f902b78c Feb 27 16:47:51 crc kubenswrapper[4814]: I0227 16:47:51.095725 4814 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 16:47:51 crc kubenswrapper[4814]: I0227 16:47:51.701972 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" event={"ID":"2bce282e-cd29-40eb-bb24-4ba8503d8a38","Type":"ContainerStarted","Data":"854534f3125369c349d87e7fb79a36b84a469a1209e020c8fcbe53e8f902b78c"} Feb 27 16:47:51 crc kubenswrapper[4814]: I0227 16:47:51.704247 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2fb1eae8-a753-43f3-8ed7-759e6e211ace","Type":"ContainerStarted","Data":"9e109ffb6502da5564ea447f0e2ab5f3cd6d394f8cb3d36a829096faa24a3251"} Feb 27 16:47:51 crc kubenswrapper[4814]: I0227 16:47:51.705365 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 27 16:47:51 crc kubenswrapper[4814]: I0227 16:47:51.745048 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.745029652 podStartE2EDuration="36.745029652s" podCreationTimestamp="2026-02-27 16:47:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:47:51.736998621 +0000 UTC m=+1484.189623451" watchObservedRunningTime="2026-02-27 16:47:51.745029652 +0000 UTC m=+1484.197654492" Feb 27 16:47:52 crc kubenswrapper[4814]: I0227 16:47:52.716583 4814 generic.go:334] "Generic (PLEG): container finished" podID="51ca92ea-753f-4e44-94a8-68e73d165193" containerID="3c171ce5597e0c2dc06c8e8eddb953dea1ee23d2a0bc2932815d090eef1b7be0" exitCode=0 Feb 27 16:47:52 crc kubenswrapper[4814]: I0227 16:47:52.716707 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"51ca92ea-753f-4e44-94a8-68e73d165193","Type":"ContainerDied","Data":"3c171ce5597e0c2dc06c8e8eddb953dea1ee23d2a0bc2932815d090eef1b7be0"} Feb 27 16:47:53 crc kubenswrapper[4814]: I0227 16:47:53.727308 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"51ca92ea-753f-4e44-94a8-68e73d165193","Type":"ContainerStarted","Data":"b7bee5c65edb489849f688897cb135850290913465da273e50a3393a1d1b7a12"} Feb 27 16:47:53 crc kubenswrapper[4814]: I0227 16:47:53.728368 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:47:53 crc kubenswrapper[4814]: I0227 16:47:53.758009 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.757991592 podStartE2EDuration="37.757991592s" podCreationTimestamp="2026-02-27 16:47:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 16:47:53.753623456 +0000 UTC m=+1486.206248286" watchObservedRunningTime="2026-02-27 16:47:53.757991592 +0000 UTC m=+1486.210616422" Feb 27 16:47:54 crc kubenswrapper[4814]: I0227 16:47:54.282102 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jsssw"] Feb 27 16:47:54 crc kubenswrapper[4814]: I0227 16:47:54.290428 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:47:54 crc kubenswrapper[4814]: I0227 16:47:54.322217 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jsssw"] Feb 27 16:47:54 crc kubenswrapper[4814]: I0227 16:47:54.387276 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-utilities\") pod \"redhat-operators-jsssw\" (UID: \"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7\") " pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:47:54 crc kubenswrapper[4814]: I0227 16:47:54.387338 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-catalog-content\") pod \"redhat-operators-jsssw\" (UID: \"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7\") " pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:47:54 crc kubenswrapper[4814]: I0227 16:47:54.387441 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl8ll\" (UniqueName: \"kubernetes.io/projected/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-kube-api-access-bl8ll\") pod \"redhat-operators-jsssw\" (UID: \"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7\") " pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:47:54 crc kubenswrapper[4814]: I0227 16:47:54.491530 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-utilities\") pod \"redhat-operators-jsssw\" (UID: \"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7\") " pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:47:54 crc kubenswrapper[4814]: I0227 16:47:54.491939 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-catalog-content\") pod \"redhat-operators-jsssw\" (UID: \"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7\") " pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:47:54 crc kubenswrapper[4814]: I0227 16:47:54.492057 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl8ll\" (UniqueName: \"kubernetes.io/projected/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-kube-api-access-bl8ll\") pod \"redhat-operators-jsssw\" (UID: \"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7\") " pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:47:54 crc kubenswrapper[4814]: I0227 16:47:54.492175 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-utilities\") pod \"redhat-operators-jsssw\" (UID: \"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7\") " pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:47:54 crc kubenswrapper[4814]: I0227 16:47:54.492364 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-catalog-content\") pod \"redhat-operators-jsssw\" (UID: \"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7\") " pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:47:54 crc kubenswrapper[4814]: I0227 16:47:54.519229 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl8ll\" (UniqueName: \"kubernetes.io/projected/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-kube-api-access-bl8ll\") pod \"redhat-operators-jsssw\" (UID: \"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7\") " pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:47:54 crc kubenswrapper[4814]: I0227 16:47:54.627876 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:48:00 crc kubenswrapper[4814]: I0227 16:48:00.172975 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536848-wt9mj"] Feb 27 16:48:00 crc kubenswrapper[4814]: I0227 16:48:00.174786 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536848-wt9mj" Feb 27 16:48:00 crc kubenswrapper[4814]: I0227 16:48:00.177053 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 16:48:00 crc kubenswrapper[4814]: I0227 16:48:00.177378 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:48:00 crc kubenswrapper[4814]: I0227 16:48:00.177958 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:48:00 crc kubenswrapper[4814]: I0227 16:48:00.200376 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536848-wt9mj"] Feb 27 16:48:00 crc kubenswrapper[4814]: I0227 16:48:00.314416 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbx8n\" (UniqueName: \"kubernetes.io/projected/e9036f38-d8ae-410f-a826-3bd8599e46ea-kube-api-access-nbx8n\") pod \"auto-csr-approver-29536848-wt9mj\" (UID: \"e9036f38-d8ae-410f-a826-3bd8599e46ea\") " pod="openshift-infra/auto-csr-approver-29536848-wt9mj" Feb 27 16:48:00 crc kubenswrapper[4814]: I0227 16:48:00.416610 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbx8n\" (UniqueName: \"kubernetes.io/projected/e9036f38-d8ae-410f-a826-3bd8599e46ea-kube-api-access-nbx8n\") pod \"auto-csr-approver-29536848-wt9mj\" (UID: \"e9036f38-d8ae-410f-a826-3bd8599e46ea\") " pod="openshift-infra/auto-csr-approver-29536848-wt9mj" Feb 27 16:48:00 crc kubenswrapper[4814]: I0227 16:48:00.434997 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbx8n\" (UniqueName: \"kubernetes.io/projected/e9036f38-d8ae-410f-a826-3bd8599e46ea-kube-api-access-nbx8n\") pod \"auto-csr-approver-29536848-wt9mj\" (UID: \"e9036f38-d8ae-410f-a826-3bd8599e46ea\") " pod="openshift-infra/auto-csr-approver-29536848-wt9mj" Feb 27 16:48:00 crc kubenswrapper[4814]: I0227 16:48:00.496094 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536848-wt9mj" Feb 27 16:48:00 crc kubenswrapper[4814]: I0227 16:48:00.797679 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" event={"ID":"2bce282e-cd29-40eb-bb24-4ba8503d8a38","Type":"ContainerStarted","Data":"86a3adeffd82cea6b4f72c6742c60f2d68eb58dfa66bba9d24332d11579fc022"} Feb 27 16:48:00 crc kubenswrapper[4814]: W0227 16:48:00.807798 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a0b95f4_ea73_4ea6_bc7a_32aa644383a7.slice/crio-9fb3fac5ebcef2c095d4d9892c323f9e6e4d425a7d6421f128674a19262a0005 WatchSource:0}: Error finding container 9fb3fac5ebcef2c095d4d9892c323f9e6e4d425a7d6421f128674a19262a0005: Status 404 returned error can't find the container with id 9fb3fac5ebcef2c095d4d9892c323f9e6e4d425a7d6421f128674a19262a0005 Feb 27 16:48:00 crc kubenswrapper[4814]: I0227 16:48:00.809475 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jsssw"] Feb 27 16:48:00 crc kubenswrapper[4814]: I0227 16:48:00.823582 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" podStartSLOduration=1.5367003449999999 podStartE2EDuration="10.823564363s" podCreationTimestamp="2026-02-27 16:47:50 +0000 UTC" firstStartedPulling="2026-02-27 16:47:51.0954845 +0000 UTC m=+1483.548109330" lastFinishedPulling="2026-02-27 16:48:00.382348498 +0000 UTC m=+1492.834973348" observedRunningTime="2026-02-27 16:48:00.816597894 +0000 UTC m=+1493.269222724" watchObservedRunningTime="2026-02-27 16:48:00.823564363 +0000 UTC m=+1493.276189183" Feb 27 16:48:00 crc kubenswrapper[4814]: W0227 16:48:00.949910 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9036f38_d8ae_410f_a826_3bd8599e46ea.slice/crio-36f95f69c34bfc9476bf1ba1683c6ced32740bda1a6de70600ed66310588d6e7 WatchSource:0}: Error finding container 36f95f69c34bfc9476bf1ba1683c6ced32740bda1a6de70600ed66310588d6e7: Status 404 returned error can't find the container with id 36f95f69c34bfc9476bf1ba1683c6ced32740bda1a6de70600ed66310588d6e7 Feb 27 16:48:00 crc kubenswrapper[4814]: I0227 16:48:00.955400 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536848-wt9mj"] Feb 27 16:48:01 crc kubenswrapper[4814]: I0227 16:48:01.811626 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536848-wt9mj" event={"ID":"e9036f38-d8ae-410f-a826-3bd8599e46ea","Type":"ContainerStarted","Data":"36f95f69c34bfc9476bf1ba1683c6ced32740bda1a6de70600ed66310588d6e7"} Feb 27 16:48:01 crc kubenswrapper[4814]: I0227 16:48:01.814866 4814 generic.go:334] "Generic (PLEG): container finished" podID="4a0b95f4-ea73-4ea6-bc7a-32aa644383a7" containerID="915b40b3e0d487af838faa986c01d14fcd654dee2e6e11862161cc4ce53596aa" exitCode=0 Feb 27 16:48:01 crc kubenswrapper[4814]: I0227 16:48:01.814926 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jsssw" event={"ID":"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7","Type":"ContainerDied","Data":"915b40b3e0d487af838faa986c01d14fcd654dee2e6e11862161cc4ce53596aa"} Feb 27 16:48:01 crc kubenswrapper[4814]: I0227 16:48:01.814973 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jsssw" event={"ID":"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7","Type":"ContainerStarted","Data":"9fb3fac5ebcef2c095d4d9892c323f9e6e4d425a7d6421f128674a19262a0005"} Feb 27 16:48:02 crc kubenswrapper[4814]: I0227 16:48:02.838127 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536848-wt9mj" event={"ID":"e9036f38-d8ae-410f-a826-3bd8599e46ea","Type":"ContainerStarted","Data":"cde917c2cc7628aa6eff69c0664fc22d8968642459b789c8e4e37989415ad0ab"} Feb 27 16:48:02 crc kubenswrapper[4814]: I0227 16:48:02.872103 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536848-wt9mj" podStartSLOduration=1.893828158 podStartE2EDuration="2.872080674s" podCreationTimestamp="2026-02-27 16:48:00 +0000 UTC" firstStartedPulling="2026-02-27 16:48:00.952167326 +0000 UTC m=+1493.404792166" lastFinishedPulling="2026-02-27 16:48:01.930419842 +0000 UTC m=+1494.383044682" observedRunningTime="2026-02-27 16:48:02.860016646 +0000 UTC m=+1495.312641496" watchObservedRunningTime="2026-02-27 16:48:02.872080674 +0000 UTC m=+1495.324705524" Feb 27 16:48:03 crc kubenswrapper[4814]: I0227 16:48:03.853511 4814 generic.go:334] "Generic (PLEG): container finished" podID="e9036f38-d8ae-410f-a826-3bd8599e46ea" containerID="cde917c2cc7628aa6eff69c0664fc22d8968642459b789c8e4e37989415ad0ab" exitCode=0 Feb 27 16:48:03 crc kubenswrapper[4814]: I0227 16:48:03.853630 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536848-wt9mj" event={"ID":"e9036f38-d8ae-410f-a826-3bd8599e46ea","Type":"ContainerDied","Data":"cde917c2cc7628aa6eff69c0664fc22d8968642459b789c8e4e37989415ad0ab"} Feb 27 16:48:03 crc kubenswrapper[4814]: I0227 16:48:03.858299 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jsssw" event={"ID":"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7","Type":"ContainerStarted","Data":"3eb2e38193d8be7989491b87bde8515fa852204863d622168683bebca2adab33"} Feb 27 16:48:04 crc kubenswrapper[4814]: I0227 16:48:04.880695 4814 generic.go:334] "Generic (PLEG): container finished" podID="4a0b95f4-ea73-4ea6-bc7a-32aa644383a7" containerID="3eb2e38193d8be7989491b87bde8515fa852204863d622168683bebca2adab33" exitCode=0 Feb 27 16:48:04 crc kubenswrapper[4814]: I0227 16:48:04.880951 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jsssw" event={"ID":"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7","Type":"ContainerDied","Data":"3eb2e38193d8be7989491b87bde8515fa852204863d622168683bebca2adab33"} Feb 27 16:48:05 crc kubenswrapper[4814]: I0227 16:48:05.287298 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536848-wt9mj" Feb 27 16:48:05 crc kubenswrapper[4814]: I0227 16:48:05.444864 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbx8n\" (UniqueName: \"kubernetes.io/projected/e9036f38-d8ae-410f-a826-3bd8599e46ea-kube-api-access-nbx8n\") pod \"e9036f38-d8ae-410f-a826-3bd8599e46ea\" (UID: \"e9036f38-d8ae-410f-a826-3bd8599e46ea\") " Feb 27 16:48:05 crc kubenswrapper[4814]: I0227 16:48:05.454494 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9036f38-d8ae-410f-a826-3bd8599e46ea-kube-api-access-nbx8n" (OuterVolumeSpecName: "kube-api-access-nbx8n") pod "e9036f38-d8ae-410f-a826-3bd8599e46ea" (UID: "e9036f38-d8ae-410f-a826-3bd8599e46ea"). InnerVolumeSpecName "kube-api-access-nbx8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:48:05 crc kubenswrapper[4814]: I0227 16:48:05.548496 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbx8n\" (UniqueName: \"kubernetes.io/projected/e9036f38-d8ae-410f-a826-3bd8599e46ea-kube-api-access-nbx8n\") on node \"crc\" DevicePath \"\"" Feb 27 16:48:05 crc kubenswrapper[4814]: I0227 16:48:05.750726 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 27 16:48:05 crc kubenswrapper[4814]: I0227 16:48:05.899168 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536848-wt9mj" event={"ID":"e9036f38-d8ae-410f-a826-3bd8599e46ea","Type":"ContainerDied","Data":"36f95f69c34bfc9476bf1ba1683c6ced32740bda1a6de70600ed66310588d6e7"} Feb 27 16:48:05 crc kubenswrapper[4814]: I0227 16:48:05.899214 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36f95f69c34bfc9476bf1ba1683c6ced32740bda1a6de70600ed66310588d6e7" Feb 27 16:48:05 crc kubenswrapper[4814]: I0227 16:48:05.899303 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536848-wt9mj" Feb 27 16:48:05 crc kubenswrapper[4814]: I0227 16:48:05.965410 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536842-hvtr9"] Feb 27 16:48:05 crc kubenswrapper[4814]: I0227 16:48:05.985875 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536842-hvtr9"] Feb 27 16:48:06 crc kubenswrapper[4814]: I0227 16:48:06.500884 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2edd190-afdf-4d53-a42e-c8a41e4a8853" path="/var/lib/kubelet/pods/c2edd190-afdf-4d53-a42e-c8a41e4a8853/volumes" Feb 27 16:48:06 crc kubenswrapper[4814]: I0227 16:48:06.905468 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 27 16:48:06 crc kubenswrapper[4814]: I0227 16:48:06.912539 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jsssw" event={"ID":"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7","Type":"ContainerStarted","Data":"e7d963f161dc7cfd6fc38b1a3f188997eac27ac8edfc42b03bd19f0d1da26224"} Feb 27 16:48:06 crc kubenswrapper[4814]: I0227 16:48:06.960838 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jsssw" podStartSLOduration=9.414085411 podStartE2EDuration="12.960807087s" podCreationTimestamp="2026-02-27 16:47:54 +0000 UTC" firstStartedPulling="2026-02-27 16:48:01.817039115 +0000 UTC m=+1494.269663945" lastFinishedPulling="2026-02-27 16:48:05.363760791 +0000 UTC m=+1497.816385621" observedRunningTime="2026-02-27 16:48:06.957514835 +0000 UTC m=+1499.410139665" watchObservedRunningTime="2026-02-27 16:48:06.960807087 +0000 UTC m=+1499.413431947" Feb 27 16:48:14 crc kubenswrapper[4814]: I0227 16:48:14.628319 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:48:14 crc kubenswrapper[4814]: I0227 16:48:14.628870 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:48:14 crc kubenswrapper[4814]: I0227 16:48:14.675792 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:48:15 crc kubenswrapper[4814]: I0227 16:48:15.072424 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:48:15 crc kubenswrapper[4814]: I0227 16:48:15.131024 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jsssw"] Feb 27 16:48:17 crc kubenswrapper[4814]: I0227 16:48:17.024896 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jsssw" podUID="4a0b95f4-ea73-4ea6-bc7a-32aa644383a7" containerName="registry-server" containerID="cri-o://e7d963f161dc7cfd6fc38b1a3f188997eac27ac8edfc42b03bd19f0d1da26224" gracePeriod=2 Feb 27 16:48:17 crc kubenswrapper[4814]: I0227 16:48:17.546492 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:48:17 crc kubenswrapper[4814]: I0227 16:48:17.597595 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl8ll\" (UniqueName: \"kubernetes.io/projected/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-kube-api-access-bl8ll\") pod \"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7\" (UID: \"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7\") " Feb 27 16:48:17 crc kubenswrapper[4814]: I0227 16:48:17.597721 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-utilities\") pod \"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7\" (UID: \"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7\") " Feb 27 16:48:17 crc kubenswrapper[4814]: I0227 16:48:17.597761 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-catalog-content\") pod \"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7\" (UID: \"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7\") " Feb 27 16:48:17 crc kubenswrapper[4814]: I0227 16:48:17.599592 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-utilities" (OuterVolumeSpecName: "utilities") pod "4a0b95f4-ea73-4ea6-bc7a-32aa644383a7" (UID: "4a0b95f4-ea73-4ea6-bc7a-32aa644383a7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:48:17 crc kubenswrapper[4814]: I0227 16:48:17.609474 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-kube-api-access-bl8ll" (OuterVolumeSpecName: "kube-api-access-bl8ll") pod "4a0b95f4-ea73-4ea6-bc7a-32aa644383a7" (UID: "4a0b95f4-ea73-4ea6-bc7a-32aa644383a7"). InnerVolumeSpecName "kube-api-access-bl8ll". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:48:17 crc kubenswrapper[4814]: I0227 16:48:17.701530 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:48:17 crc kubenswrapper[4814]: I0227 16:48:17.701584 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl8ll\" (UniqueName: \"kubernetes.io/projected/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-kube-api-access-bl8ll\") on node \"crc\" DevicePath \"\"" Feb 27 16:48:17 crc kubenswrapper[4814]: I0227 16:48:17.716062 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4a0b95f4-ea73-4ea6-bc7a-32aa644383a7" (UID: "4a0b95f4-ea73-4ea6-bc7a-32aa644383a7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:48:17 crc kubenswrapper[4814]: I0227 16:48:17.803111 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:48:18 crc kubenswrapper[4814]: I0227 16:48:18.035431 4814 generic.go:334] "Generic (PLEG): container finished" podID="4a0b95f4-ea73-4ea6-bc7a-32aa644383a7" containerID="e7d963f161dc7cfd6fc38b1a3f188997eac27ac8edfc42b03bd19f0d1da26224" exitCode=0 Feb 27 16:48:18 crc kubenswrapper[4814]: I0227 16:48:18.035474 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jsssw" event={"ID":"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7","Type":"ContainerDied","Data":"e7d963f161dc7cfd6fc38b1a3f188997eac27ac8edfc42b03bd19f0d1da26224"} Feb 27 16:48:18 crc kubenswrapper[4814]: I0227 16:48:18.035487 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jsssw" Feb 27 16:48:18 crc kubenswrapper[4814]: I0227 16:48:18.035504 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jsssw" event={"ID":"4a0b95f4-ea73-4ea6-bc7a-32aa644383a7","Type":"ContainerDied","Data":"9fb3fac5ebcef2c095d4d9892c323f9e6e4d425a7d6421f128674a19262a0005"} Feb 27 16:48:18 crc kubenswrapper[4814]: I0227 16:48:18.035523 4814 scope.go:117] "RemoveContainer" containerID="e7d963f161dc7cfd6fc38b1a3f188997eac27ac8edfc42b03bd19f0d1da26224" Feb 27 16:48:18 crc kubenswrapper[4814]: I0227 16:48:18.070640 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jsssw"] Feb 27 16:48:18 crc kubenswrapper[4814]: I0227 16:48:18.070746 4814 scope.go:117] "RemoveContainer" containerID="3eb2e38193d8be7989491b87bde8515fa852204863d622168683bebca2adab33" Feb 27 16:48:18 crc kubenswrapper[4814]: I0227 16:48:18.079106 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jsssw"] Feb 27 16:48:18 crc kubenswrapper[4814]: I0227 16:48:18.097167 4814 scope.go:117] "RemoveContainer" containerID="915b40b3e0d487af838faa986c01d14fcd654dee2e6e11862161cc4ce53596aa" Feb 27 16:48:18 crc kubenswrapper[4814]: I0227 16:48:18.173200 4814 scope.go:117] "RemoveContainer" containerID="e7d963f161dc7cfd6fc38b1a3f188997eac27ac8edfc42b03bd19f0d1da26224" Feb 27 16:48:18 crc kubenswrapper[4814]: E0227 16:48:18.173768 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7d963f161dc7cfd6fc38b1a3f188997eac27ac8edfc42b03bd19f0d1da26224\": container with ID starting with e7d963f161dc7cfd6fc38b1a3f188997eac27ac8edfc42b03bd19f0d1da26224 not found: ID does not exist" containerID="e7d963f161dc7cfd6fc38b1a3f188997eac27ac8edfc42b03bd19f0d1da26224" Feb 27 16:48:18 crc kubenswrapper[4814]: I0227 16:48:18.173815 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7d963f161dc7cfd6fc38b1a3f188997eac27ac8edfc42b03bd19f0d1da26224"} err="failed to get container status \"e7d963f161dc7cfd6fc38b1a3f188997eac27ac8edfc42b03bd19f0d1da26224\": rpc error: code = NotFound desc = could not find container \"e7d963f161dc7cfd6fc38b1a3f188997eac27ac8edfc42b03bd19f0d1da26224\": container with ID starting with e7d963f161dc7cfd6fc38b1a3f188997eac27ac8edfc42b03bd19f0d1da26224 not found: ID does not exist" Feb 27 16:48:18 crc kubenswrapper[4814]: I0227 16:48:18.173845 4814 scope.go:117] "RemoveContainer" containerID="3eb2e38193d8be7989491b87bde8515fa852204863d622168683bebca2adab33" Feb 27 16:48:18 crc kubenswrapper[4814]: E0227 16:48:18.174262 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eb2e38193d8be7989491b87bde8515fa852204863d622168683bebca2adab33\": container with ID starting with 3eb2e38193d8be7989491b87bde8515fa852204863d622168683bebca2adab33 not found: ID does not exist" containerID="3eb2e38193d8be7989491b87bde8515fa852204863d622168683bebca2adab33" Feb 27 16:48:18 crc kubenswrapper[4814]: I0227 16:48:18.174305 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eb2e38193d8be7989491b87bde8515fa852204863d622168683bebca2adab33"} err="failed to get container status \"3eb2e38193d8be7989491b87bde8515fa852204863d622168683bebca2adab33\": rpc error: code = NotFound desc = could not find container \"3eb2e38193d8be7989491b87bde8515fa852204863d622168683bebca2adab33\": container with ID starting with 3eb2e38193d8be7989491b87bde8515fa852204863d622168683bebca2adab33 not found: ID does not exist" Feb 27 16:48:18 crc kubenswrapper[4814]: I0227 16:48:18.174321 4814 scope.go:117] "RemoveContainer" containerID="915b40b3e0d487af838faa986c01d14fcd654dee2e6e11862161cc4ce53596aa" Feb 27 16:48:18 crc kubenswrapper[4814]: E0227 16:48:18.174757 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"915b40b3e0d487af838faa986c01d14fcd654dee2e6e11862161cc4ce53596aa\": container with ID starting with 915b40b3e0d487af838faa986c01d14fcd654dee2e6e11862161cc4ce53596aa not found: ID does not exist" containerID="915b40b3e0d487af838faa986c01d14fcd654dee2e6e11862161cc4ce53596aa" Feb 27 16:48:18 crc kubenswrapper[4814]: I0227 16:48:18.174785 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"915b40b3e0d487af838faa986c01d14fcd654dee2e6e11862161cc4ce53596aa"} err="failed to get container status \"915b40b3e0d487af838faa986c01d14fcd654dee2e6e11862161cc4ce53596aa\": rpc error: code = NotFound desc = could not find container \"915b40b3e0d487af838faa986c01d14fcd654dee2e6e11862161cc4ce53596aa\": container with ID starting with 915b40b3e0d487af838faa986c01d14fcd654dee2e6e11862161cc4ce53596aa not found: ID does not exist" Feb 27 16:48:18 crc kubenswrapper[4814]: I0227 16:48:18.509798 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a0b95f4-ea73-4ea6-bc7a-32aa644383a7" path="/var/lib/kubelet/pods/4a0b95f4-ea73-4ea6-bc7a-32aa644383a7/volumes" Feb 27 16:48:22 crc kubenswrapper[4814]: I0227 16:48:22.902864 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:48:22 crc kubenswrapper[4814]: I0227 16:48:22.903312 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:48:23 crc kubenswrapper[4814]: I0227 16:48:23.099892 4814 generic.go:334] "Generic (PLEG): container finished" podID="2bce282e-cd29-40eb-bb24-4ba8503d8a38" containerID="86a3adeffd82cea6b4f72c6742c60f2d68eb58dfa66bba9d24332d11579fc022" exitCode=0 Feb 27 16:48:23 crc kubenswrapper[4814]: I0227 16:48:23.099965 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" event={"ID":"2bce282e-cd29-40eb-bb24-4ba8503d8a38","Type":"ContainerDied","Data":"86a3adeffd82cea6b4f72c6742c60f2d68eb58dfa66bba9d24332d11579fc022"} Feb 27 16:48:24 crc kubenswrapper[4814]: I0227 16:48:24.625812 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" Feb 27 16:48:24 crc kubenswrapper[4814]: I0227 16:48:24.663751 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-ssh-key-openstack-edpm-ipam\") pod \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\" (UID: \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\") " Feb 27 16:48:24 crc kubenswrapper[4814]: I0227 16:48:24.664030 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-inventory\") pod \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\" (UID: \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\") " Feb 27 16:48:24 crc kubenswrapper[4814]: I0227 16:48:24.664078 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vczvq\" (UniqueName: \"kubernetes.io/projected/2bce282e-cd29-40eb-bb24-4ba8503d8a38-kube-api-access-vczvq\") pod \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\" (UID: \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\") " Feb 27 16:48:24 crc kubenswrapper[4814]: I0227 16:48:24.664140 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-repo-setup-combined-ca-bundle\") pod \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\" (UID: \"2bce282e-cd29-40eb-bb24-4ba8503d8a38\") " Feb 27 16:48:24 crc kubenswrapper[4814]: I0227 16:48:24.673134 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "2bce282e-cd29-40eb-bb24-4ba8503d8a38" (UID: "2bce282e-cd29-40eb-bb24-4ba8503d8a38"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:48:24 crc kubenswrapper[4814]: I0227 16:48:24.674416 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bce282e-cd29-40eb-bb24-4ba8503d8a38-kube-api-access-vczvq" (OuterVolumeSpecName: "kube-api-access-vczvq") pod "2bce282e-cd29-40eb-bb24-4ba8503d8a38" (UID: "2bce282e-cd29-40eb-bb24-4ba8503d8a38"). InnerVolumeSpecName "kube-api-access-vczvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:48:24 crc kubenswrapper[4814]: I0227 16:48:24.693545 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-inventory" (OuterVolumeSpecName: "inventory") pod "2bce282e-cd29-40eb-bb24-4ba8503d8a38" (UID: "2bce282e-cd29-40eb-bb24-4ba8503d8a38"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:48:24 crc kubenswrapper[4814]: I0227 16:48:24.695630 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2bce282e-cd29-40eb-bb24-4ba8503d8a38" (UID: "2bce282e-cd29-40eb-bb24-4ba8503d8a38"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:48:24 crc kubenswrapper[4814]: I0227 16:48:24.766362 4814 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 16:48:24 crc kubenswrapper[4814]: I0227 16:48:24.766397 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vczvq\" (UniqueName: \"kubernetes.io/projected/2bce282e-cd29-40eb-bb24-4ba8503d8a38-kube-api-access-vczvq\") on node \"crc\" DevicePath \"\"" Feb 27 16:48:24 crc kubenswrapper[4814]: I0227 16:48:24.766407 4814 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:48:24 crc kubenswrapper[4814]: I0227 16:48:24.766417 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2bce282e-cd29-40eb-bb24-4ba8503d8a38-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.126308 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" event={"ID":"2bce282e-cd29-40eb-bb24-4ba8503d8a38","Type":"ContainerDied","Data":"854534f3125369c349d87e7fb79a36b84a469a1209e020c8fcbe53e8f902b78c"} Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.126757 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="854534f3125369c349d87e7fb79a36b84a469a1209e020c8fcbe53e8f902b78c" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.126397 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.287472 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq"] Feb 27 16:48:25 crc kubenswrapper[4814]: E0227 16:48:25.288313 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a0b95f4-ea73-4ea6-bc7a-32aa644383a7" containerName="extract-utilities" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.288329 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a0b95f4-ea73-4ea6-bc7a-32aa644383a7" containerName="extract-utilities" Feb 27 16:48:25 crc kubenswrapper[4814]: E0227 16:48:25.288350 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a0b95f4-ea73-4ea6-bc7a-32aa644383a7" containerName="extract-content" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.288358 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a0b95f4-ea73-4ea6-bc7a-32aa644383a7" containerName="extract-content" Feb 27 16:48:25 crc kubenswrapper[4814]: E0227 16:48:25.288384 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bce282e-cd29-40eb-bb24-4ba8503d8a38" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.288394 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bce282e-cd29-40eb-bb24-4ba8503d8a38" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 27 16:48:25 crc kubenswrapper[4814]: E0227 16:48:25.288418 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a0b95f4-ea73-4ea6-bc7a-32aa644383a7" containerName="registry-server" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.288425 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a0b95f4-ea73-4ea6-bc7a-32aa644383a7" containerName="registry-server" Feb 27 16:48:25 crc kubenswrapper[4814]: E0227 16:48:25.288441 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9036f38-d8ae-410f-a826-3bd8599e46ea" containerName="oc" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.288449 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9036f38-d8ae-410f-a826-3bd8599e46ea" containerName="oc" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.288667 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9036f38-d8ae-410f-a826-3bd8599e46ea" containerName="oc" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.288683 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bce282e-cd29-40eb-bb24-4ba8503d8a38" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.288708 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a0b95f4-ea73-4ea6-bc7a-32aa644383a7" containerName="registry-server" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.289510 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.293666 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.293684 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.293941 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.294017 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.298896 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq"] Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.376930 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4d3976db-1cd1-4f9b-8975-038c57210a39-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rqbpq\" (UID: \"4d3976db-1cd1-4f9b-8975-038c57210a39\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.377013 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d3976db-1cd1-4f9b-8975-038c57210a39-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rqbpq\" (UID: \"4d3976db-1cd1-4f9b-8975-038c57210a39\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.377122 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89q24\" (UniqueName: \"kubernetes.io/projected/4d3976db-1cd1-4f9b-8975-038c57210a39-kube-api-access-89q24\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rqbpq\" (UID: \"4d3976db-1cd1-4f9b-8975-038c57210a39\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.479215 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89q24\" (UniqueName: \"kubernetes.io/projected/4d3976db-1cd1-4f9b-8975-038c57210a39-kube-api-access-89q24\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rqbpq\" (UID: \"4d3976db-1cd1-4f9b-8975-038c57210a39\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.479310 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4d3976db-1cd1-4f9b-8975-038c57210a39-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rqbpq\" (UID: \"4d3976db-1cd1-4f9b-8975-038c57210a39\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.479390 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d3976db-1cd1-4f9b-8975-038c57210a39-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rqbpq\" (UID: \"4d3976db-1cd1-4f9b-8975-038c57210a39\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.489422 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4d3976db-1cd1-4f9b-8975-038c57210a39-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rqbpq\" (UID: \"4d3976db-1cd1-4f9b-8975-038c57210a39\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.491916 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d3976db-1cd1-4f9b-8975-038c57210a39-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rqbpq\" (UID: \"4d3976db-1cd1-4f9b-8975-038c57210a39\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.510897 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89q24\" (UniqueName: \"kubernetes.io/projected/4d3976db-1cd1-4f9b-8975-038c57210a39-kube-api-access-89q24\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rqbpq\" (UID: \"4d3976db-1cd1-4f9b-8975-038c57210a39\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" Feb 27 16:48:25 crc kubenswrapper[4814]: I0227 16:48:25.634520 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" Feb 27 16:48:26 crc kubenswrapper[4814]: I0227 16:48:26.185199 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq"] Feb 27 16:48:27 crc kubenswrapper[4814]: I0227 16:48:27.160246 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" event={"ID":"4d3976db-1cd1-4f9b-8975-038c57210a39","Type":"ContainerStarted","Data":"ac4fcf8e69b74ba19860fdb8b3cf2824c21f07d8930442a8e80844b8fd5073c9"} Feb 27 16:48:27 crc kubenswrapper[4814]: I0227 16:48:27.160750 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" event={"ID":"4d3976db-1cd1-4f9b-8975-038c57210a39","Type":"ContainerStarted","Data":"11009dfc6878a052dae27f5a07f832c152518cb96a2154d3770374df8970f621"} Feb 27 16:48:27 crc kubenswrapper[4814]: I0227 16:48:27.188982 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" podStartSLOduration=1.596031874 podStartE2EDuration="2.188954204s" podCreationTimestamp="2026-02-27 16:48:25 +0000 UTC" firstStartedPulling="2026-02-27 16:48:26.191851778 +0000 UTC m=+1518.644476608" lastFinishedPulling="2026-02-27 16:48:26.784774088 +0000 UTC m=+1519.237398938" observedRunningTime="2026-02-27 16:48:27.186093755 +0000 UTC m=+1519.638718625" watchObservedRunningTime="2026-02-27 16:48:27.188954204 +0000 UTC m=+1519.641579064" Feb 27 16:48:30 crc kubenswrapper[4814]: I0227 16:48:30.200888 4814 generic.go:334] "Generic (PLEG): container finished" podID="4d3976db-1cd1-4f9b-8975-038c57210a39" containerID="ac4fcf8e69b74ba19860fdb8b3cf2824c21f07d8930442a8e80844b8fd5073c9" exitCode=0 Feb 27 16:48:30 crc kubenswrapper[4814]: I0227 16:48:30.201017 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" event={"ID":"4d3976db-1cd1-4f9b-8975-038c57210a39","Type":"ContainerDied","Data":"ac4fcf8e69b74ba19860fdb8b3cf2824c21f07d8930442a8e80844b8fd5073c9"} Feb 27 16:48:31 crc kubenswrapper[4814]: I0227 16:48:31.615842 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" Feb 27 16:48:31 crc kubenswrapper[4814]: I0227 16:48:31.621783 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89q24\" (UniqueName: \"kubernetes.io/projected/4d3976db-1cd1-4f9b-8975-038c57210a39-kube-api-access-89q24\") pod \"4d3976db-1cd1-4f9b-8975-038c57210a39\" (UID: \"4d3976db-1cd1-4f9b-8975-038c57210a39\") " Feb 27 16:48:31 crc kubenswrapper[4814]: I0227 16:48:31.621853 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d3976db-1cd1-4f9b-8975-038c57210a39-inventory\") pod \"4d3976db-1cd1-4f9b-8975-038c57210a39\" (UID: \"4d3976db-1cd1-4f9b-8975-038c57210a39\") " Feb 27 16:48:31 crc kubenswrapper[4814]: I0227 16:48:31.621923 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4d3976db-1cd1-4f9b-8975-038c57210a39-ssh-key-openstack-edpm-ipam\") pod \"4d3976db-1cd1-4f9b-8975-038c57210a39\" (UID: \"4d3976db-1cd1-4f9b-8975-038c57210a39\") " Feb 27 16:48:31 crc kubenswrapper[4814]: I0227 16:48:31.629047 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d3976db-1cd1-4f9b-8975-038c57210a39-kube-api-access-89q24" (OuterVolumeSpecName: "kube-api-access-89q24") pod "4d3976db-1cd1-4f9b-8975-038c57210a39" (UID: "4d3976db-1cd1-4f9b-8975-038c57210a39"). InnerVolumeSpecName "kube-api-access-89q24". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:48:31 crc kubenswrapper[4814]: I0227 16:48:31.658347 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d3976db-1cd1-4f9b-8975-038c57210a39-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4d3976db-1cd1-4f9b-8975-038c57210a39" (UID: "4d3976db-1cd1-4f9b-8975-038c57210a39"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:48:31 crc kubenswrapper[4814]: I0227 16:48:31.662152 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d3976db-1cd1-4f9b-8975-038c57210a39-inventory" (OuterVolumeSpecName: "inventory") pod "4d3976db-1cd1-4f9b-8975-038c57210a39" (UID: "4d3976db-1cd1-4f9b-8975-038c57210a39"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:48:31 crc kubenswrapper[4814]: I0227 16:48:31.724079 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89q24\" (UniqueName: \"kubernetes.io/projected/4d3976db-1cd1-4f9b-8975-038c57210a39-kube-api-access-89q24\") on node \"crc\" DevicePath \"\"" Feb 27 16:48:31 crc kubenswrapper[4814]: I0227 16:48:31.724109 4814 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d3976db-1cd1-4f9b-8975-038c57210a39-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 16:48:31 crc kubenswrapper[4814]: I0227 16:48:31.724120 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4d3976db-1cd1-4f9b-8975-038c57210a39-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.225937 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" event={"ID":"4d3976db-1cd1-4f9b-8975-038c57210a39","Type":"ContainerDied","Data":"11009dfc6878a052dae27f5a07f832c152518cb96a2154d3770374df8970f621"} Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.226008 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11009dfc6878a052dae27f5a07f832c152518cb96a2154d3770374df8970f621" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.226032 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rqbpq" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.311028 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs"] Feb 27 16:48:32 crc kubenswrapper[4814]: E0227 16:48:32.311463 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d3976db-1cd1-4f9b-8975-038c57210a39" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.311480 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d3976db-1cd1-4f9b-8975-038c57210a39" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.311670 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d3976db-1cd1-4f9b-8975-038c57210a39" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.312299 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.314202 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.314568 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.315478 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.315722 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.319898 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs"] Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.340384 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs\" (UID: \"556cb206-711d-4002-80cf-1ffe3b8f9643\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.340461 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8vj2\" (UniqueName: \"kubernetes.io/projected/556cb206-711d-4002-80cf-1ffe3b8f9643-kube-api-access-d8vj2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs\" (UID: \"556cb206-711d-4002-80cf-1ffe3b8f9643\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.340526 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs\" (UID: \"556cb206-711d-4002-80cf-1ffe3b8f9643\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.340630 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs\" (UID: \"556cb206-711d-4002-80cf-1ffe3b8f9643\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.442902 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs\" (UID: \"556cb206-711d-4002-80cf-1ffe3b8f9643\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.442997 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8vj2\" (UniqueName: \"kubernetes.io/projected/556cb206-711d-4002-80cf-1ffe3b8f9643-kube-api-access-d8vj2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs\" (UID: \"556cb206-711d-4002-80cf-1ffe3b8f9643\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.443060 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs\" (UID: \"556cb206-711d-4002-80cf-1ffe3b8f9643\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.443135 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs\" (UID: \"556cb206-711d-4002-80cf-1ffe3b8f9643\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.448351 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs\" (UID: \"556cb206-711d-4002-80cf-1ffe3b8f9643\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.455284 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs\" (UID: \"556cb206-711d-4002-80cf-1ffe3b8f9643\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.455729 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs\" (UID: \"556cb206-711d-4002-80cf-1ffe3b8f9643\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.461233 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8vj2\" (UniqueName: \"kubernetes.io/projected/556cb206-711d-4002-80cf-1ffe3b8f9643-kube-api-access-d8vj2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs\" (UID: \"556cb206-711d-4002-80cf-1ffe3b8f9643\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" Feb 27 16:48:32 crc kubenswrapper[4814]: I0227 16:48:32.635329 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" Feb 27 16:48:33 crc kubenswrapper[4814]: I0227 16:48:33.304810 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs"] Feb 27 16:48:33 crc kubenswrapper[4814]: W0227 16:48:33.305163 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod556cb206_711d_4002_80cf_1ffe3b8f9643.slice/crio-df5f8105fdb4661ebb70ca502b1375d7393b60b2e2f298b91d2cded6702d5bd3 WatchSource:0}: Error finding container df5f8105fdb4661ebb70ca502b1375d7393b60b2e2f298b91d2cded6702d5bd3: Status 404 returned error can't find the container with id df5f8105fdb4661ebb70ca502b1375d7393b60b2e2f298b91d2cded6702d5bd3 Feb 27 16:48:34 crc kubenswrapper[4814]: I0227 16:48:34.260834 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" event={"ID":"556cb206-711d-4002-80cf-1ffe3b8f9643","Type":"ContainerStarted","Data":"c51a37fc066a6e05a93414b59974b5a7fa7857ca94dab9d0e3f03c9b28517eb5"} Feb 27 16:48:34 crc kubenswrapper[4814]: I0227 16:48:34.261110 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" event={"ID":"556cb206-711d-4002-80cf-1ffe3b8f9643","Type":"ContainerStarted","Data":"df5f8105fdb4661ebb70ca502b1375d7393b60b2e2f298b91d2cded6702d5bd3"} Feb 27 16:48:37 crc kubenswrapper[4814]: I0227 16:48:37.486071 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" podStartSLOduration=5.007213098 podStartE2EDuration="5.486045319s" podCreationTimestamp="2026-02-27 16:48:32 +0000 UTC" firstStartedPulling="2026-02-27 16:48:33.308906279 +0000 UTC m=+1525.761531149" lastFinishedPulling="2026-02-27 16:48:33.78773854 +0000 UTC m=+1526.240363370" observedRunningTime="2026-02-27 16:48:34.286897838 +0000 UTC m=+1526.739522668" watchObservedRunningTime="2026-02-27 16:48:37.486045319 +0000 UTC m=+1529.938670169" Feb 27 16:48:37 crc kubenswrapper[4814]: I0227 16:48:37.502535 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v97zn"] Feb 27 16:48:37 crc kubenswrapper[4814]: I0227 16:48:37.508010 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:37 crc kubenswrapper[4814]: I0227 16:48:37.516586 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v97zn"] Feb 27 16:48:37 crc kubenswrapper[4814]: I0227 16:48:37.581308 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jdfd\" (UniqueName: \"kubernetes.io/projected/5be77f28-08c7-4121-8a2c-634839625d9c-kube-api-access-8jdfd\") pod \"redhat-marketplace-v97zn\" (UID: \"5be77f28-08c7-4121-8a2c-634839625d9c\") " pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:37 crc kubenswrapper[4814]: I0227 16:48:37.581389 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5be77f28-08c7-4121-8a2c-634839625d9c-utilities\") pod \"redhat-marketplace-v97zn\" (UID: \"5be77f28-08c7-4121-8a2c-634839625d9c\") " pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:37 crc kubenswrapper[4814]: I0227 16:48:37.581456 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5be77f28-08c7-4121-8a2c-634839625d9c-catalog-content\") pod \"redhat-marketplace-v97zn\" (UID: \"5be77f28-08c7-4121-8a2c-634839625d9c\") " pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:37 crc kubenswrapper[4814]: I0227 16:48:37.684170 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jdfd\" (UniqueName: \"kubernetes.io/projected/5be77f28-08c7-4121-8a2c-634839625d9c-kube-api-access-8jdfd\") pod \"redhat-marketplace-v97zn\" (UID: \"5be77f28-08c7-4121-8a2c-634839625d9c\") " pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:37 crc kubenswrapper[4814]: I0227 16:48:37.684220 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5be77f28-08c7-4121-8a2c-634839625d9c-utilities\") pod \"redhat-marketplace-v97zn\" (UID: \"5be77f28-08c7-4121-8a2c-634839625d9c\") " pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:37 crc kubenswrapper[4814]: I0227 16:48:37.684283 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5be77f28-08c7-4121-8a2c-634839625d9c-catalog-content\") pod \"redhat-marketplace-v97zn\" (UID: \"5be77f28-08c7-4121-8a2c-634839625d9c\") " pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:37 crc kubenswrapper[4814]: I0227 16:48:37.685326 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5be77f28-08c7-4121-8a2c-634839625d9c-catalog-content\") pod \"redhat-marketplace-v97zn\" (UID: \"5be77f28-08c7-4121-8a2c-634839625d9c\") " pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:37 crc kubenswrapper[4814]: I0227 16:48:37.685342 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5be77f28-08c7-4121-8a2c-634839625d9c-utilities\") pod \"redhat-marketplace-v97zn\" (UID: \"5be77f28-08c7-4121-8a2c-634839625d9c\") " pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:37 crc kubenswrapper[4814]: I0227 16:48:37.708144 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jdfd\" (UniqueName: \"kubernetes.io/projected/5be77f28-08c7-4121-8a2c-634839625d9c-kube-api-access-8jdfd\") pod \"redhat-marketplace-v97zn\" (UID: \"5be77f28-08c7-4121-8a2c-634839625d9c\") " pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:37 crc kubenswrapper[4814]: I0227 16:48:37.833554 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:38 crc kubenswrapper[4814]: I0227 16:48:38.334376 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v97zn"] Feb 27 16:48:39 crc kubenswrapper[4814]: I0227 16:48:39.310337 4814 generic.go:334] "Generic (PLEG): container finished" podID="5be77f28-08c7-4121-8a2c-634839625d9c" containerID="a853a0de3b034ba2c7e5d25630655f5e9d6e738f48237e4b07e3d18ae978c9ed" exitCode=0 Feb 27 16:48:39 crc kubenswrapper[4814]: I0227 16:48:39.310402 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v97zn" event={"ID":"5be77f28-08c7-4121-8a2c-634839625d9c","Type":"ContainerDied","Data":"a853a0de3b034ba2c7e5d25630655f5e9d6e738f48237e4b07e3d18ae978c9ed"} Feb 27 16:48:39 crc kubenswrapper[4814]: I0227 16:48:39.310441 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v97zn" event={"ID":"5be77f28-08c7-4121-8a2c-634839625d9c","Type":"ContainerStarted","Data":"a2b9976d6aba01a7b62b8c73b4aff4794272f1177d9dcff761622b263e3b8185"} Feb 27 16:48:40 crc kubenswrapper[4814]: I0227 16:48:40.323612 4814 generic.go:334] "Generic (PLEG): container finished" podID="5be77f28-08c7-4121-8a2c-634839625d9c" containerID="f937a3489c038dabc74868cf7cb1d09b992406d1d746fa40426e0aa3d4bcc766" exitCode=0 Feb 27 16:48:40 crc kubenswrapper[4814]: I0227 16:48:40.323671 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v97zn" event={"ID":"5be77f28-08c7-4121-8a2c-634839625d9c","Type":"ContainerDied","Data":"f937a3489c038dabc74868cf7cb1d09b992406d1d746fa40426e0aa3d4bcc766"} Feb 27 16:48:41 crc kubenswrapper[4814]: I0227 16:48:41.373780 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v97zn" event={"ID":"5be77f28-08c7-4121-8a2c-634839625d9c","Type":"ContainerStarted","Data":"9c5c1d5ac00c4cbfda0d7e15c8070334d75d156494e66d4ffcc0cfbbd46c7c73"} Feb 27 16:48:41 crc kubenswrapper[4814]: I0227 16:48:41.432647 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v97zn" podStartSLOduration=2.992997823 podStartE2EDuration="4.432622725s" podCreationTimestamp="2026-02-27 16:48:37 +0000 UTC" firstStartedPulling="2026-02-27 16:48:39.31592348 +0000 UTC m=+1531.768548310" lastFinishedPulling="2026-02-27 16:48:40.755548372 +0000 UTC m=+1533.208173212" observedRunningTime="2026-02-27 16:48:41.42028275 +0000 UTC m=+1533.872907580" watchObservedRunningTime="2026-02-27 16:48:41.432622725 +0000 UTC m=+1533.885247555" Feb 27 16:48:47 crc kubenswrapper[4814]: I0227 16:48:47.834736 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:47 crc kubenswrapper[4814]: I0227 16:48:47.835250 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:47 crc kubenswrapper[4814]: I0227 16:48:47.946351 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:48 crc kubenswrapper[4814]: I0227 16:48:48.553013 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:48 crc kubenswrapper[4814]: I0227 16:48:48.626220 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v97zn"] Feb 27 16:48:50 crc kubenswrapper[4814]: I0227 16:48:50.480470 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v97zn" podUID="5be77f28-08c7-4121-8a2c-634839625d9c" containerName="registry-server" containerID="cri-o://9c5c1d5ac00c4cbfda0d7e15c8070334d75d156494e66d4ffcc0cfbbd46c7c73" gracePeriod=2 Feb 27 16:48:50 crc kubenswrapper[4814]: I0227 16:48:50.983680 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.085499 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5be77f28-08c7-4121-8a2c-634839625d9c-catalog-content\") pod \"5be77f28-08c7-4121-8a2c-634839625d9c\" (UID: \"5be77f28-08c7-4121-8a2c-634839625d9c\") " Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.085742 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jdfd\" (UniqueName: \"kubernetes.io/projected/5be77f28-08c7-4121-8a2c-634839625d9c-kube-api-access-8jdfd\") pod \"5be77f28-08c7-4121-8a2c-634839625d9c\" (UID: \"5be77f28-08c7-4121-8a2c-634839625d9c\") " Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.085891 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5be77f28-08c7-4121-8a2c-634839625d9c-utilities\") pod \"5be77f28-08c7-4121-8a2c-634839625d9c\" (UID: \"5be77f28-08c7-4121-8a2c-634839625d9c\") " Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.086934 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5be77f28-08c7-4121-8a2c-634839625d9c-utilities" (OuterVolumeSpecName: "utilities") pod "5be77f28-08c7-4121-8a2c-634839625d9c" (UID: "5be77f28-08c7-4121-8a2c-634839625d9c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.096126 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5be77f28-08c7-4121-8a2c-634839625d9c-kube-api-access-8jdfd" (OuterVolumeSpecName: "kube-api-access-8jdfd") pod "5be77f28-08c7-4121-8a2c-634839625d9c" (UID: "5be77f28-08c7-4121-8a2c-634839625d9c"). InnerVolumeSpecName "kube-api-access-8jdfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.122718 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5be77f28-08c7-4121-8a2c-634839625d9c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5be77f28-08c7-4121-8a2c-634839625d9c" (UID: "5be77f28-08c7-4121-8a2c-634839625d9c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.187781 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5be77f28-08c7-4121-8a2c-634839625d9c-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.187809 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5be77f28-08c7-4121-8a2c-634839625d9c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.187821 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jdfd\" (UniqueName: \"kubernetes.io/projected/5be77f28-08c7-4121-8a2c-634839625d9c-kube-api-access-8jdfd\") on node \"crc\" DevicePath \"\"" Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.494903 4814 generic.go:334] "Generic (PLEG): container finished" podID="5be77f28-08c7-4121-8a2c-634839625d9c" containerID="9c5c1d5ac00c4cbfda0d7e15c8070334d75d156494e66d4ffcc0cfbbd46c7c73" exitCode=0 Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.494974 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v97zn" Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.494992 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v97zn" event={"ID":"5be77f28-08c7-4121-8a2c-634839625d9c","Type":"ContainerDied","Data":"9c5c1d5ac00c4cbfda0d7e15c8070334d75d156494e66d4ffcc0cfbbd46c7c73"} Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.495071 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v97zn" event={"ID":"5be77f28-08c7-4121-8a2c-634839625d9c","Type":"ContainerDied","Data":"a2b9976d6aba01a7b62b8c73b4aff4794272f1177d9dcff761622b263e3b8185"} Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.495105 4814 scope.go:117] "RemoveContainer" containerID="9c5c1d5ac00c4cbfda0d7e15c8070334d75d156494e66d4ffcc0cfbbd46c7c73" Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.548156 4814 scope.go:117] "RemoveContainer" containerID="f937a3489c038dabc74868cf7cb1d09b992406d1d746fa40426e0aa3d4bcc766" Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.551317 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v97zn"] Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.565897 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v97zn"] Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.579120 4814 scope.go:117] "RemoveContainer" containerID="a853a0de3b034ba2c7e5d25630655f5e9d6e738f48237e4b07e3d18ae978c9ed" Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.632081 4814 scope.go:117] "RemoveContainer" containerID="9c5c1d5ac00c4cbfda0d7e15c8070334d75d156494e66d4ffcc0cfbbd46c7c73" Feb 27 16:48:51 crc kubenswrapper[4814]: E0227 16:48:51.632633 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c5c1d5ac00c4cbfda0d7e15c8070334d75d156494e66d4ffcc0cfbbd46c7c73\": container with ID starting with 9c5c1d5ac00c4cbfda0d7e15c8070334d75d156494e66d4ffcc0cfbbd46c7c73 not found: ID does not exist" containerID="9c5c1d5ac00c4cbfda0d7e15c8070334d75d156494e66d4ffcc0cfbbd46c7c73" Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.632680 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c5c1d5ac00c4cbfda0d7e15c8070334d75d156494e66d4ffcc0cfbbd46c7c73"} err="failed to get container status \"9c5c1d5ac00c4cbfda0d7e15c8070334d75d156494e66d4ffcc0cfbbd46c7c73\": rpc error: code = NotFound desc = could not find container \"9c5c1d5ac00c4cbfda0d7e15c8070334d75d156494e66d4ffcc0cfbbd46c7c73\": container with ID starting with 9c5c1d5ac00c4cbfda0d7e15c8070334d75d156494e66d4ffcc0cfbbd46c7c73 not found: ID does not exist" Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.632706 4814 scope.go:117] "RemoveContainer" containerID="f937a3489c038dabc74868cf7cb1d09b992406d1d746fa40426e0aa3d4bcc766" Feb 27 16:48:51 crc kubenswrapper[4814]: E0227 16:48:51.633000 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f937a3489c038dabc74868cf7cb1d09b992406d1d746fa40426e0aa3d4bcc766\": container with ID starting with f937a3489c038dabc74868cf7cb1d09b992406d1d746fa40426e0aa3d4bcc766 not found: ID does not exist" containerID="f937a3489c038dabc74868cf7cb1d09b992406d1d746fa40426e0aa3d4bcc766" Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.633035 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f937a3489c038dabc74868cf7cb1d09b992406d1d746fa40426e0aa3d4bcc766"} err="failed to get container status \"f937a3489c038dabc74868cf7cb1d09b992406d1d746fa40426e0aa3d4bcc766\": rpc error: code = NotFound desc = could not find container \"f937a3489c038dabc74868cf7cb1d09b992406d1d746fa40426e0aa3d4bcc766\": container with ID starting with f937a3489c038dabc74868cf7cb1d09b992406d1d746fa40426e0aa3d4bcc766 not found: ID does not exist" Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.633054 4814 scope.go:117] "RemoveContainer" containerID="a853a0de3b034ba2c7e5d25630655f5e9d6e738f48237e4b07e3d18ae978c9ed" Feb 27 16:48:51 crc kubenswrapper[4814]: E0227 16:48:51.633603 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a853a0de3b034ba2c7e5d25630655f5e9d6e738f48237e4b07e3d18ae978c9ed\": container with ID starting with a853a0de3b034ba2c7e5d25630655f5e9d6e738f48237e4b07e3d18ae978c9ed not found: ID does not exist" containerID="a853a0de3b034ba2c7e5d25630655f5e9d6e738f48237e4b07e3d18ae978c9ed" Feb 27 16:48:51 crc kubenswrapper[4814]: I0227 16:48:51.633624 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a853a0de3b034ba2c7e5d25630655f5e9d6e738f48237e4b07e3d18ae978c9ed"} err="failed to get container status \"a853a0de3b034ba2c7e5d25630655f5e9d6e738f48237e4b07e3d18ae978c9ed\": rpc error: code = NotFound desc = could not find container \"a853a0de3b034ba2c7e5d25630655f5e9d6e738f48237e4b07e3d18ae978c9ed\": container with ID starting with a853a0de3b034ba2c7e5d25630655f5e9d6e738f48237e4b07e3d18ae978c9ed not found: ID does not exist" Feb 27 16:48:52 crc kubenswrapper[4814]: I0227 16:48:52.514057 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5be77f28-08c7-4121-8a2c-634839625d9c" path="/var/lib/kubelet/pods/5be77f28-08c7-4121-8a2c-634839625d9c/volumes" Feb 27 16:48:52 crc kubenswrapper[4814]: I0227 16:48:52.903088 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:48:52 crc kubenswrapper[4814]: I0227 16:48:52.903188 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:48:53 crc kubenswrapper[4814]: I0227 16:48:53.164483 4814 scope.go:117] "RemoveContainer" containerID="224f2879b4b05aa596bc464639c7088cf0190a24804cdf9cadaff1d692aae3b3" Feb 27 16:48:53 crc kubenswrapper[4814]: I0227 16:48:53.201368 4814 scope.go:117] "RemoveContainer" containerID="e02cda0f2dd0eec50057588dc8a38645b05d52a2e967dd1110002b8ccbab6273" Feb 27 16:48:53 crc kubenswrapper[4814]: I0227 16:48:53.290069 4814 scope.go:117] "RemoveContainer" containerID="ba75ebdc4923da74d710b0b62d701e35a031dd4a66ec36db84223e7fdbd063ea" Feb 27 16:48:53 crc kubenswrapper[4814]: I0227 16:48:53.325560 4814 scope.go:117] "RemoveContainer" containerID="91632a104e79e38caed145351cf91b5ac043517bc37f9849b25e7a6cbb580785" Feb 27 16:49:22 crc kubenswrapper[4814]: I0227 16:49:22.903167 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:49:22 crc kubenswrapper[4814]: I0227 16:49:22.904035 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:49:22 crc kubenswrapper[4814]: I0227 16:49:22.904101 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:49:22 crc kubenswrapper[4814]: I0227 16:49:22.905133 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e5df7c66a9ca17c8ac1eb1c76463afbe46a665cbf347524e84122f10413b151c"} pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 16:49:22 crc kubenswrapper[4814]: I0227 16:49:22.905237 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" containerID="cri-o://e5df7c66a9ca17c8ac1eb1c76463afbe46a665cbf347524e84122f10413b151c" gracePeriod=600 Feb 27 16:49:23 crc kubenswrapper[4814]: I0227 16:49:23.862982 4814 generic.go:334] "Generic (PLEG): container finished" podID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerID="e5df7c66a9ca17c8ac1eb1c76463afbe46a665cbf347524e84122f10413b151c" exitCode=0 Feb 27 16:49:23 crc kubenswrapper[4814]: I0227 16:49:23.863076 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerDied","Data":"e5df7c66a9ca17c8ac1eb1c76463afbe46a665cbf347524e84122f10413b151c"} Feb 27 16:49:23 crc kubenswrapper[4814]: I0227 16:49:23.863631 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08"} Feb 27 16:49:23 crc kubenswrapper[4814]: I0227 16:49:23.863661 4814 scope.go:117] "RemoveContainer" containerID="5e8afb4996159ac03d9ce86a713b7b4aae36d378d653cf752342458cdea88224" Feb 27 16:49:53 crc kubenswrapper[4814]: I0227 16:49:53.491086 4814 scope.go:117] "RemoveContainer" containerID="eb9a34f2ce8980b3b94acf9e587c2bde662f1e06e113cc61530af63123fca3fa" Feb 27 16:49:53 crc kubenswrapper[4814]: I0227 16:49:53.536061 4814 scope.go:117] "RemoveContainer" containerID="868dca07dd0ef20b6d4dab49ac197397d1fe8947c8339c8c88f4eaf0da2a6e69" Feb 27 16:49:53 crc kubenswrapper[4814]: I0227 16:49:53.571691 4814 scope.go:117] "RemoveContainer" containerID="49a68508357e578f6c9b99dce6fa55364fe996ac9b6d9f43c57b556c3b745576" Feb 27 16:49:53 crc kubenswrapper[4814]: I0227 16:49:53.624299 4814 scope.go:117] "RemoveContainer" containerID="c0e2764da868d4e939f9b18a6abbd8e1879216738d1cd0705ff76e8814168a53" Feb 27 16:49:53 crc kubenswrapper[4814]: I0227 16:49:53.681898 4814 scope.go:117] "RemoveContainer" containerID="ba8e4d12b3a356f51ad9bcd93b46a3983571ebb22168d59fa156bd0aa103a861" Feb 27 16:49:53 crc kubenswrapper[4814]: I0227 16:49:53.737486 4814 scope.go:117] "RemoveContainer" containerID="25b206576573c6339f705f0fbee3821f6900b276d6dbecd7dce3be15bc5f87d3" Feb 27 16:49:53 crc kubenswrapper[4814]: I0227 16:49:53.770517 4814 scope.go:117] "RemoveContainer" containerID="0f868ed4357392cf34291c9a8d69618903599c078e3052e5c429cb032301b99b" Feb 27 16:49:53 crc kubenswrapper[4814]: I0227 16:49:53.826390 4814 scope.go:117] "RemoveContainer" containerID="dc548cb7d34cbc790dcb1a9abab7241efa6d0ba3b73528769736a659769b0128" Feb 27 16:49:53 crc kubenswrapper[4814]: I0227 16:49:53.894507 4814 scope.go:117] "RemoveContainer" containerID="c52d992512b39f9030e3eb2b20244de9ced5f2493ef47f17365e81f5d87b6cf2" Feb 27 16:50:00 crc kubenswrapper[4814]: I0227 16:50:00.150105 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536850-xqntl"] Feb 27 16:50:00 crc kubenswrapper[4814]: E0227 16:50:00.151018 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5be77f28-08c7-4121-8a2c-634839625d9c" containerName="extract-content" Feb 27 16:50:00 crc kubenswrapper[4814]: I0227 16:50:00.151033 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="5be77f28-08c7-4121-8a2c-634839625d9c" containerName="extract-content" Feb 27 16:50:00 crc kubenswrapper[4814]: E0227 16:50:00.151054 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5be77f28-08c7-4121-8a2c-634839625d9c" containerName="registry-server" Feb 27 16:50:00 crc kubenswrapper[4814]: I0227 16:50:00.151062 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="5be77f28-08c7-4121-8a2c-634839625d9c" containerName="registry-server" Feb 27 16:50:00 crc kubenswrapper[4814]: E0227 16:50:00.151094 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5be77f28-08c7-4121-8a2c-634839625d9c" containerName="extract-utilities" Feb 27 16:50:00 crc kubenswrapper[4814]: I0227 16:50:00.151103 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="5be77f28-08c7-4121-8a2c-634839625d9c" containerName="extract-utilities" Feb 27 16:50:00 crc kubenswrapper[4814]: I0227 16:50:00.151355 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="5be77f28-08c7-4121-8a2c-634839625d9c" containerName="registry-server" Feb 27 16:50:00 crc kubenswrapper[4814]: I0227 16:50:00.152064 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536850-xqntl" Feb 27 16:50:00 crc kubenswrapper[4814]: I0227 16:50:00.155422 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 16:50:00 crc kubenswrapper[4814]: I0227 16:50:00.155956 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:50:00 crc kubenswrapper[4814]: I0227 16:50:00.156090 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:50:00 crc kubenswrapper[4814]: I0227 16:50:00.179458 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536850-xqntl"] Feb 27 16:50:00 crc kubenswrapper[4814]: I0227 16:50:00.312236 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njtgz\" (UniqueName: \"kubernetes.io/projected/b661ed9f-ec8e-42bd-86eb-3c4da181ced4-kube-api-access-njtgz\") pod \"auto-csr-approver-29536850-xqntl\" (UID: \"b661ed9f-ec8e-42bd-86eb-3c4da181ced4\") " pod="openshift-infra/auto-csr-approver-29536850-xqntl" Feb 27 16:50:00 crc kubenswrapper[4814]: I0227 16:50:00.414048 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njtgz\" (UniqueName: \"kubernetes.io/projected/b661ed9f-ec8e-42bd-86eb-3c4da181ced4-kube-api-access-njtgz\") pod \"auto-csr-approver-29536850-xqntl\" (UID: \"b661ed9f-ec8e-42bd-86eb-3c4da181ced4\") " pod="openshift-infra/auto-csr-approver-29536850-xqntl" Feb 27 16:50:00 crc kubenswrapper[4814]: I0227 16:50:00.434065 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njtgz\" (UniqueName: \"kubernetes.io/projected/b661ed9f-ec8e-42bd-86eb-3c4da181ced4-kube-api-access-njtgz\") pod \"auto-csr-approver-29536850-xqntl\" (UID: \"b661ed9f-ec8e-42bd-86eb-3c4da181ced4\") " pod="openshift-infra/auto-csr-approver-29536850-xqntl" Feb 27 16:50:00 crc kubenswrapper[4814]: I0227 16:50:00.478809 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536850-xqntl" Feb 27 16:50:01 crc kubenswrapper[4814]: I0227 16:50:01.027771 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536850-xqntl"] Feb 27 16:50:01 crc kubenswrapper[4814]: W0227 16:50:01.034603 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb661ed9f_ec8e_42bd_86eb_3c4da181ced4.slice/crio-63ed51ac84d1d9828b870b027e9690336da733e6bd220b7d796f938478a3c12b WatchSource:0}: Error finding container 63ed51ac84d1d9828b870b027e9690336da733e6bd220b7d796f938478a3c12b: Status 404 returned error can't find the container with id 63ed51ac84d1d9828b870b027e9690336da733e6bd220b7d796f938478a3c12b Feb 27 16:50:01 crc kubenswrapper[4814]: I0227 16:50:01.339842 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536850-xqntl" event={"ID":"b661ed9f-ec8e-42bd-86eb-3c4da181ced4","Type":"ContainerStarted","Data":"63ed51ac84d1d9828b870b027e9690336da733e6bd220b7d796f938478a3c12b"} Feb 27 16:50:03 crc kubenswrapper[4814]: I0227 16:50:03.360196 4814 generic.go:334] "Generic (PLEG): container finished" podID="b661ed9f-ec8e-42bd-86eb-3c4da181ced4" containerID="c7592a56dad2fc0610cf175f91deea1aaf285a476965cc1103a83e5c84d3d6cf" exitCode=0 Feb 27 16:50:03 crc kubenswrapper[4814]: I0227 16:50:03.360303 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536850-xqntl" event={"ID":"b661ed9f-ec8e-42bd-86eb-3c4da181ced4","Type":"ContainerDied","Data":"c7592a56dad2fc0610cf175f91deea1aaf285a476965cc1103a83e5c84d3d6cf"} Feb 27 16:50:04 crc kubenswrapper[4814]: I0227 16:50:04.772947 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536850-xqntl" Feb 27 16:50:04 crc kubenswrapper[4814]: I0227 16:50:04.913217 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njtgz\" (UniqueName: \"kubernetes.io/projected/b661ed9f-ec8e-42bd-86eb-3c4da181ced4-kube-api-access-njtgz\") pod \"b661ed9f-ec8e-42bd-86eb-3c4da181ced4\" (UID: \"b661ed9f-ec8e-42bd-86eb-3c4da181ced4\") " Feb 27 16:50:04 crc kubenswrapper[4814]: I0227 16:50:04.924726 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b661ed9f-ec8e-42bd-86eb-3c4da181ced4-kube-api-access-njtgz" (OuterVolumeSpecName: "kube-api-access-njtgz") pod "b661ed9f-ec8e-42bd-86eb-3c4da181ced4" (UID: "b661ed9f-ec8e-42bd-86eb-3c4da181ced4"). InnerVolumeSpecName "kube-api-access-njtgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:50:05 crc kubenswrapper[4814]: I0227 16:50:05.015458 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njtgz\" (UniqueName: \"kubernetes.io/projected/b661ed9f-ec8e-42bd-86eb-3c4da181ced4-kube-api-access-njtgz\") on node \"crc\" DevicePath \"\"" Feb 27 16:50:05 crc kubenswrapper[4814]: I0227 16:50:05.386919 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536850-xqntl" event={"ID":"b661ed9f-ec8e-42bd-86eb-3c4da181ced4","Type":"ContainerDied","Data":"63ed51ac84d1d9828b870b027e9690336da733e6bd220b7d796f938478a3c12b"} Feb 27 16:50:05 crc kubenswrapper[4814]: I0227 16:50:05.386973 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63ed51ac84d1d9828b870b027e9690336da733e6bd220b7d796f938478a3c12b" Feb 27 16:50:05 crc kubenswrapper[4814]: I0227 16:50:05.387064 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536850-xqntl" Feb 27 16:50:05 crc kubenswrapper[4814]: I0227 16:50:05.879151 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536844-4sfb5"] Feb 27 16:50:05 crc kubenswrapper[4814]: I0227 16:50:05.894281 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536844-4sfb5"] Feb 27 16:50:06 crc kubenswrapper[4814]: I0227 16:50:06.502075 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68c41339-6536-4998-901e-179aee23ac8c" path="/var/lib/kubelet/pods/68c41339-6536-4998-901e-179aee23ac8c/volumes" Feb 27 16:50:43 crc kubenswrapper[4814]: I0227 16:50:43.069618 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qz6kc"] Feb 27 16:50:43 crc kubenswrapper[4814]: E0227 16:50:43.070918 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b661ed9f-ec8e-42bd-86eb-3c4da181ced4" containerName="oc" Feb 27 16:50:43 crc kubenswrapper[4814]: I0227 16:50:43.070940 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="b661ed9f-ec8e-42bd-86eb-3c4da181ced4" containerName="oc" Feb 27 16:50:43 crc kubenswrapper[4814]: I0227 16:50:43.071321 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="b661ed9f-ec8e-42bd-86eb-3c4da181ced4" containerName="oc" Feb 27 16:50:43 crc kubenswrapper[4814]: I0227 16:50:43.073725 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:43 crc kubenswrapper[4814]: I0227 16:50:43.089670 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qz6kc"] Feb 27 16:50:43 crc kubenswrapper[4814]: I0227 16:50:43.166991 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j2c8\" (UniqueName: \"kubernetes.io/projected/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-kube-api-access-8j2c8\") pod \"certified-operators-qz6kc\" (UID: \"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09\") " pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:43 crc kubenswrapper[4814]: I0227 16:50:43.167118 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-catalog-content\") pod \"certified-operators-qz6kc\" (UID: \"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09\") " pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:43 crc kubenswrapper[4814]: I0227 16:50:43.167541 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-utilities\") pod \"certified-operators-qz6kc\" (UID: \"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09\") " pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:43 crc kubenswrapper[4814]: I0227 16:50:43.269504 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-catalog-content\") pod \"certified-operators-qz6kc\" (UID: \"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09\") " pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:43 crc kubenswrapper[4814]: I0227 16:50:43.269652 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-utilities\") pod \"certified-operators-qz6kc\" (UID: \"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09\") " pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:43 crc kubenswrapper[4814]: I0227 16:50:43.269703 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j2c8\" (UniqueName: \"kubernetes.io/projected/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-kube-api-access-8j2c8\") pod \"certified-operators-qz6kc\" (UID: \"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09\") " pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:43 crc kubenswrapper[4814]: I0227 16:50:43.270066 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-catalog-content\") pod \"certified-operators-qz6kc\" (UID: \"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09\") " pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:43 crc kubenswrapper[4814]: I0227 16:50:43.270150 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-utilities\") pod \"certified-operators-qz6kc\" (UID: \"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09\") " pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:43 crc kubenswrapper[4814]: I0227 16:50:43.298375 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j2c8\" (UniqueName: \"kubernetes.io/projected/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-kube-api-access-8j2c8\") pod \"certified-operators-qz6kc\" (UID: \"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09\") " pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:43 crc kubenswrapper[4814]: I0227 16:50:43.417109 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:43 crc kubenswrapper[4814]: I0227 16:50:43.943167 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qz6kc"] Feb 27 16:50:44 crc kubenswrapper[4814]: I0227 16:50:44.925164 4814 generic.go:334] "Generic (PLEG): container finished" podID="bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09" containerID="b7f4e61daaa874899b61761593bff37a5f0c01638cb066229e626099c1adf188" exitCode=0 Feb 27 16:50:44 crc kubenswrapper[4814]: I0227 16:50:44.925412 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz6kc" event={"ID":"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09","Type":"ContainerDied","Data":"b7f4e61daaa874899b61761593bff37a5f0c01638cb066229e626099c1adf188"} Feb 27 16:50:44 crc kubenswrapper[4814]: I0227 16:50:44.925825 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz6kc" event={"ID":"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09","Type":"ContainerStarted","Data":"0cf8b1cdf904e38f60d60c8bfe1f1478bd3a5409d584106fe3f92bf2cdb158a7"} Feb 27 16:50:45 crc kubenswrapper[4814]: I0227 16:50:45.939537 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz6kc" event={"ID":"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09","Type":"ContainerStarted","Data":"92f5d0105b3426331baa6d7d6be68b937819fa6a976aba064a3af0a99afd322c"} Feb 27 16:50:46 crc kubenswrapper[4814]: I0227 16:50:46.955957 4814 generic.go:334] "Generic (PLEG): container finished" podID="bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09" containerID="92f5d0105b3426331baa6d7d6be68b937819fa6a976aba064a3af0a99afd322c" exitCode=0 Feb 27 16:50:46 crc kubenswrapper[4814]: I0227 16:50:46.956036 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz6kc" event={"ID":"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09","Type":"ContainerDied","Data":"92f5d0105b3426331baa6d7d6be68b937819fa6a976aba064a3af0a99afd322c"} Feb 27 16:50:47 crc kubenswrapper[4814]: I0227 16:50:47.967903 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz6kc" event={"ID":"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09","Type":"ContainerStarted","Data":"f004442b58a9c3ae17e947ea8d74d5a7c776b037cec1be0fe663bc8f96c4f9da"} Feb 27 16:50:48 crc kubenswrapper[4814]: I0227 16:50:48.002578 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qz6kc" podStartSLOduration=2.537476813 podStartE2EDuration="5.002552828s" podCreationTimestamp="2026-02-27 16:50:43 +0000 UTC" firstStartedPulling="2026-02-27 16:50:44.929176488 +0000 UTC m=+1657.381801358" lastFinishedPulling="2026-02-27 16:50:47.394252543 +0000 UTC m=+1659.846877373" observedRunningTime="2026-02-27 16:50:47.99077019 +0000 UTC m=+1660.443395050" watchObservedRunningTime="2026-02-27 16:50:48.002552828 +0000 UTC m=+1660.455177688" Feb 27 16:50:53 crc kubenswrapper[4814]: I0227 16:50:53.417747 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:53 crc kubenswrapper[4814]: I0227 16:50:53.418365 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:53 crc kubenswrapper[4814]: I0227 16:50:53.468607 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:54 crc kubenswrapper[4814]: I0227 16:50:54.008873 4814 scope.go:117] "RemoveContainer" containerID="6eb8cd24d369d2d009c45dc31fe5e0ecd17dd117cae0420b6edc105714bbbd82" Feb 27 16:50:54 crc kubenswrapper[4814]: I0227 16:50:54.078724 4814 scope.go:117] "RemoveContainer" containerID="146c26d9df94e91bd53b52abca4f6443179c2c8e2365581e20b12612c0129aca" Feb 27 16:50:54 crc kubenswrapper[4814]: I0227 16:50:54.110857 4814 scope.go:117] "RemoveContainer" containerID="fe992a34a5bd33198fa76c1d351deb8c6a8cb1d10babbc60cf65c741734db762" Feb 27 16:50:54 crc kubenswrapper[4814]: I0227 16:50:54.118521 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:54 crc kubenswrapper[4814]: I0227 16:50:54.190425 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qz6kc"] Feb 27 16:50:56 crc kubenswrapper[4814]: I0227 16:50:56.061778 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qz6kc" podUID="bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09" containerName="registry-server" containerID="cri-o://f004442b58a9c3ae17e947ea8d74d5a7c776b037cec1be0fe663bc8f96c4f9da" gracePeriod=2 Feb 27 16:50:56 crc kubenswrapper[4814]: I0227 16:50:56.629372 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:56 crc kubenswrapper[4814]: I0227 16:50:56.711651 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8j2c8\" (UniqueName: \"kubernetes.io/projected/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-kube-api-access-8j2c8\") pod \"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09\" (UID: \"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09\") " Feb 27 16:50:56 crc kubenswrapper[4814]: I0227 16:50:56.711839 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-catalog-content\") pod \"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09\" (UID: \"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09\") " Feb 27 16:50:56 crc kubenswrapper[4814]: I0227 16:50:56.712141 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-utilities\") pod \"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09\" (UID: \"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09\") " Feb 27 16:50:56 crc kubenswrapper[4814]: I0227 16:50:56.713774 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-utilities" (OuterVolumeSpecName: "utilities") pod "bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09" (UID: "bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:50:56 crc kubenswrapper[4814]: I0227 16:50:56.714475 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:50:56 crc kubenswrapper[4814]: I0227 16:50:56.723633 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-kube-api-access-8j2c8" (OuterVolumeSpecName: "kube-api-access-8j2c8") pod "bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09" (UID: "bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09"). InnerVolumeSpecName "kube-api-access-8j2c8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:50:56 crc kubenswrapper[4814]: I0227 16:50:56.817162 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8j2c8\" (UniqueName: \"kubernetes.io/projected/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-kube-api-access-8j2c8\") on node \"crc\" DevicePath \"\"" Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.079054 4814 generic.go:334] "Generic (PLEG): container finished" podID="bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09" containerID="f004442b58a9c3ae17e947ea8d74d5a7c776b037cec1be0fe663bc8f96c4f9da" exitCode=0 Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.079120 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz6kc" event={"ID":"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09","Type":"ContainerDied","Data":"f004442b58a9c3ae17e947ea8d74d5a7c776b037cec1be0fe663bc8f96c4f9da"} Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.079162 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz6kc" event={"ID":"bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09","Type":"ContainerDied","Data":"0cf8b1cdf904e38f60d60c8bfe1f1478bd3a5409d584106fe3f92bf2cdb158a7"} Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.079247 4814 scope.go:117] "RemoveContainer" containerID="f004442b58a9c3ae17e947ea8d74d5a7c776b037cec1be0fe663bc8f96c4f9da" Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.079476 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qz6kc" Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.124898 4814 scope.go:117] "RemoveContainer" containerID="92f5d0105b3426331baa6d7d6be68b937819fa6a976aba064a3af0a99afd322c" Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.135326 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09" (UID: "bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.160448 4814 scope.go:117] "RemoveContainer" containerID="b7f4e61daaa874899b61761593bff37a5f0c01638cb066229e626099c1adf188" Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.227849 4814 scope.go:117] "RemoveContainer" containerID="f004442b58a9c3ae17e947ea8d74d5a7c776b037cec1be0fe663bc8f96c4f9da" Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.229166 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:50:57 crc kubenswrapper[4814]: E0227 16:50:57.232485 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f004442b58a9c3ae17e947ea8d74d5a7c776b037cec1be0fe663bc8f96c4f9da\": container with ID starting with f004442b58a9c3ae17e947ea8d74d5a7c776b037cec1be0fe663bc8f96c4f9da not found: ID does not exist" containerID="f004442b58a9c3ae17e947ea8d74d5a7c776b037cec1be0fe663bc8f96c4f9da" Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.232543 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f004442b58a9c3ae17e947ea8d74d5a7c776b037cec1be0fe663bc8f96c4f9da"} err="failed to get container status \"f004442b58a9c3ae17e947ea8d74d5a7c776b037cec1be0fe663bc8f96c4f9da\": rpc error: code = NotFound desc = could not find container \"f004442b58a9c3ae17e947ea8d74d5a7c776b037cec1be0fe663bc8f96c4f9da\": container with ID starting with f004442b58a9c3ae17e947ea8d74d5a7c776b037cec1be0fe663bc8f96c4f9da not found: ID does not exist" Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.232578 4814 scope.go:117] "RemoveContainer" containerID="92f5d0105b3426331baa6d7d6be68b937819fa6a976aba064a3af0a99afd322c" Feb 27 16:50:57 crc kubenswrapper[4814]: E0227 16:50:57.233140 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92f5d0105b3426331baa6d7d6be68b937819fa6a976aba064a3af0a99afd322c\": container with ID starting with 92f5d0105b3426331baa6d7d6be68b937819fa6a976aba064a3af0a99afd322c not found: ID does not exist" containerID="92f5d0105b3426331baa6d7d6be68b937819fa6a976aba064a3af0a99afd322c" Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.233392 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92f5d0105b3426331baa6d7d6be68b937819fa6a976aba064a3af0a99afd322c"} err="failed to get container status \"92f5d0105b3426331baa6d7d6be68b937819fa6a976aba064a3af0a99afd322c\": rpc error: code = NotFound desc = could not find container \"92f5d0105b3426331baa6d7d6be68b937819fa6a976aba064a3af0a99afd322c\": container with ID starting with 92f5d0105b3426331baa6d7d6be68b937819fa6a976aba064a3af0a99afd322c not found: ID does not exist" Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.233566 4814 scope.go:117] "RemoveContainer" containerID="b7f4e61daaa874899b61761593bff37a5f0c01638cb066229e626099c1adf188" Feb 27 16:50:57 crc kubenswrapper[4814]: E0227 16:50:57.234358 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7f4e61daaa874899b61761593bff37a5f0c01638cb066229e626099c1adf188\": container with ID starting with b7f4e61daaa874899b61761593bff37a5f0c01638cb066229e626099c1adf188 not found: ID does not exist" containerID="b7f4e61daaa874899b61761593bff37a5f0c01638cb066229e626099c1adf188" Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.234564 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7f4e61daaa874899b61761593bff37a5f0c01638cb066229e626099c1adf188"} err="failed to get container status \"b7f4e61daaa874899b61761593bff37a5f0c01638cb066229e626099c1adf188\": rpc error: code = NotFound desc = could not find container \"b7f4e61daaa874899b61761593bff37a5f0c01638cb066229e626099c1adf188\": container with ID starting with b7f4e61daaa874899b61761593bff37a5f0c01638cb066229e626099c1adf188 not found: ID does not exist" Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.489781 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qz6kc"] Feb 27 16:50:57 crc kubenswrapper[4814]: I0227 16:50:57.504373 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qz6kc"] Feb 27 16:50:58 crc kubenswrapper[4814]: I0227 16:50:58.506084 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09" path="/var/lib/kubelet/pods/bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09/volumes" Feb 27 16:51:34 crc kubenswrapper[4814]: I0227 16:51:34.553112 4814 generic.go:334] "Generic (PLEG): container finished" podID="556cb206-711d-4002-80cf-1ffe3b8f9643" containerID="c51a37fc066a6e05a93414b59974b5a7fa7857ca94dab9d0e3f03c9b28517eb5" exitCode=0 Feb 27 16:51:34 crc kubenswrapper[4814]: I0227 16:51:34.553733 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" event={"ID":"556cb206-711d-4002-80cf-1ffe3b8f9643","Type":"ContainerDied","Data":"c51a37fc066a6e05a93414b59974b5a7fa7857ca94dab9d0e3f03c9b28517eb5"} Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.051156 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.061086 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-ssh-key-openstack-edpm-ipam\") pod \"556cb206-711d-4002-80cf-1ffe3b8f9643\" (UID: \"556cb206-711d-4002-80cf-1ffe3b8f9643\") " Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.061121 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8vj2\" (UniqueName: \"kubernetes.io/projected/556cb206-711d-4002-80cf-1ffe3b8f9643-kube-api-access-d8vj2\") pod \"556cb206-711d-4002-80cf-1ffe3b8f9643\" (UID: \"556cb206-711d-4002-80cf-1ffe3b8f9643\") " Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.061157 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-inventory\") pod \"556cb206-711d-4002-80cf-1ffe3b8f9643\" (UID: \"556cb206-711d-4002-80cf-1ffe3b8f9643\") " Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.061218 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-bootstrap-combined-ca-bundle\") pod \"556cb206-711d-4002-80cf-1ffe3b8f9643\" (UID: \"556cb206-711d-4002-80cf-1ffe3b8f9643\") " Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.073988 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/556cb206-711d-4002-80cf-1ffe3b8f9643-kube-api-access-d8vj2" (OuterVolumeSpecName: "kube-api-access-d8vj2") pod "556cb206-711d-4002-80cf-1ffe3b8f9643" (UID: "556cb206-711d-4002-80cf-1ffe3b8f9643"). InnerVolumeSpecName "kube-api-access-d8vj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.094239 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "556cb206-711d-4002-80cf-1ffe3b8f9643" (UID: "556cb206-711d-4002-80cf-1ffe3b8f9643"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.104377 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-inventory" (OuterVolumeSpecName: "inventory") pod "556cb206-711d-4002-80cf-1ffe3b8f9643" (UID: "556cb206-711d-4002-80cf-1ffe3b8f9643"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.109951 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "556cb206-711d-4002-80cf-1ffe3b8f9643" (UID: "556cb206-711d-4002-80cf-1ffe3b8f9643"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.164393 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.164635 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8vj2\" (UniqueName: \"kubernetes.io/projected/556cb206-711d-4002-80cf-1ffe3b8f9643-kube-api-access-d8vj2\") on node \"crc\" DevicePath \"\"" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.164655 4814 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.164668 4814 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/556cb206-711d-4002-80cf-1ffe3b8f9643-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.576361 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" event={"ID":"556cb206-711d-4002-80cf-1ffe3b8f9643","Type":"ContainerDied","Data":"df5f8105fdb4661ebb70ca502b1375d7393b60b2e2f298b91d2cded6702d5bd3"} Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.576399 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df5f8105fdb4661ebb70ca502b1375d7393b60b2e2f298b91d2cded6702d5bd3" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.576450 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.682011 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq"] Feb 27 16:51:36 crc kubenswrapper[4814]: E0227 16:51:36.682580 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09" containerName="extract-content" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.682613 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09" containerName="extract-content" Feb 27 16:51:36 crc kubenswrapper[4814]: E0227 16:51:36.682640 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09" containerName="registry-server" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.682650 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09" containerName="registry-server" Feb 27 16:51:36 crc kubenswrapper[4814]: E0227 16:51:36.682699 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09" containerName="extract-utilities" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.682710 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09" containerName="extract-utilities" Feb 27 16:51:36 crc kubenswrapper[4814]: E0227 16:51:36.682739 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="556cb206-711d-4002-80cf-1ffe3b8f9643" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.682752 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="556cb206-711d-4002-80cf-1ffe3b8f9643" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.683031 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="556cb206-711d-4002-80cf-1ffe3b8f9643" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.683079 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdd8fde4-d0d9-4d3a-a21f-2f25e6136d09" containerName="registry-server" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.684175 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.689719 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.689863 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.689927 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.691119 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.699626 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq"] Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.776299 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q22z\" (UniqueName: \"kubernetes.io/projected/daf9ab93-de26-4929-a827-925d1650af5a-kube-api-access-7q22z\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq\" (UID: \"daf9ab93-de26-4929-a827-925d1650af5a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.776482 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/daf9ab93-de26-4929-a827-925d1650af5a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq\" (UID: \"daf9ab93-de26-4929-a827-925d1650af5a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.776615 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/daf9ab93-de26-4929-a827-925d1650af5a-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq\" (UID: \"daf9ab93-de26-4929-a827-925d1650af5a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.878544 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/daf9ab93-de26-4929-a827-925d1650af5a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq\" (UID: \"daf9ab93-de26-4929-a827-925d1650af5a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.878653 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/daf9ab93-de26-4929-a827-925d1650af5a-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq\" (UID: \"daf9ab93-de26-4929-a827-925d1650af5a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.878786 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q22z\" (UniqueName: \"kubernetes.io/projected/daf9ab93-de26-4929-a827-925d1650af5a-kube-api-access-7q22z\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq\" (UID: \"daf9ab93-de26-4929-a827-925d1650af5a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.885642 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/daf9ab93-de26-4929-a827-925d1650af5a-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq\" (UID: \"daf9ab93-de26-4929-a827-925d1650af5a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.889937 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/daf9ab93-de26-4929-a827-925d1650af5a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq\" (UID: \"daf9ab93-de26-4929-a827-925d1650af5a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" Feb 27 16:51:36 crc kubenswrapper[4814]: I0227 16:51:36.900711 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q22z\" (UniqueName: \"kubernetes.io/projected/daf9ab93-de26-4929-a827-925d1650af5a-kube-api-access-7q22z\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq\" (UID: \"daf9ab93-de26-4929-a827-925d1650af5a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" Feb 27 16:51:37 crc kubenswrapper[4814]: I0227 16:51:37.017455 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" Feb 27 16:51:37 crc kubenswrapper[4814]: I0227 16:51:37.548017 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq"] Feb 27 16:51:37 crc kubenswrapper[4814]: I0227 16:51:37.594533 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" event={"ID":"daf9ab93-de26-4929-a827-925d1650af5a","Type":"ContainerStarted","Data":"7d5080a90ba10a4e5b8b6af1710068b374ad6588da17fa89a73a2a013cc46815"} Feb 27 16:51:38 crc kubenswrapper[4814]: I0227 16:51:38.606154 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" event={"ID":"daf9ab93-de26-4929-a827-925d1650af5a","Type":"ContainerStarted","Data":"ac1eaf9f77f22aaa93155580aa1c7879f4e0d7724be68589bb5830957e266a8f"} Feb 27 16:51:38 crc kubenswrapper[4814]: I0227 16:51:38.627701 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" podStartSLOduration=2.067710743 podStartE2EDuration="2.627679421s" podCreationTimestamp="2026-02-27 16:51:36 +0000 UTC" firstStartedPulling="2026-02-27 16:51:37.543057786 +0000 UTC m=+1709.995682636" lastFinishedPulling="2026-02-27 16:51:38.103026454 +0000 UTC m=+1710.555651314" observedRunningTime="2026-02-27 16:51:38.626615798 +0000 UTC m=+1711.079240668" watchObservedRunningTime="2026-02-27 16:51:38.627679421 +0000 UTC m=+1711.080304251" Feb 27 16:51:52 crc kubenswrapper[4814]: I0227 16:51:52.902870 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:51:52 crc kubenswrapper[4814]: I0227 16:51:52.903607 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:52:00 crc kubenswrapper[4814]: I0227 16:52:00.166307 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536852-8vpsm"] Feb 27 16:52:00 crc kubenswrapper[4814]: I0227 16:52:00.168363 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536852-8vpsm" Feb 27 16:52:00 crc kubenswrapper[4814]: I0227 16:52:00.171467 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:52:00 crc kubenswrapper[4814]: I0227 16:52:00.171851 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:52:00 crc kubenswrapper[4814]: I0227 16:52:00.172095 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 16:52:00 crc kubenswrapper[4814]: I0227 16:52:00.186937 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536852-8vpsm"] Feb 27 16:52:00 crc kubenswrapper[4814]: I0227 16:52:00.256919 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xbc5\" (UniqueName: \"kubernetes.io/projected/93b95ff2-824a-43ff-9110-ef73f1fcc34c-kube-api-access-9xbc5\") pod \"auto-csr-approver-29536852-8vpsm\" (UID: \"93b95ff2-824a-43ff-9110-ef73f1fcc34c\") " pod="openshift-infra/auto-csr-approver-29536852-8vpsm" Feb 27 16:52:00 crc kubenswrapper[4814]: I0227 16:52:00.360907 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xbc5\" (UniqueName: \"kubernetes.io/projected/93b95ff2-824a-43ff-9110-ef73f1fcc34c-kube-api-access-9xbc5\") pod \"auto-csr-approver-29536852-8vpsm\" (UID: \"93b95ff2-824a-43ff-9110-ef73f1fcc34c\") " pod="openshift-infra/auto-csr-approver-29536852-8vpsm" Feb 27 16:52:00 crc kubenswrapper[4814]: I0227 16:52:00.394059 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xbc5\" (UniqueName: \"kubernetes.io/projected/93b95ff2-824a-43ff-9110-ef73f1fcc34c-kube-api-access-9xbc5\") pod \"auto-csr-approver-29536852-8vpsm\" (UID: \"93b95ff2-824a-43ff-9110-ef73f1fcc34c\") " pod="openshift-infra/auto-csr-approver-29536852-8vpsm" Feb 27 16:52:00 crc kubenswrapper[4814]: I0227 16:52:00.499945 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536852-8vpsm" Feb 27 16:52:01 crc kubenswrapper[4814]: W0227 16:52:01.011945 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93b95ff2_824a_43ff_9110_ef73f1fcc34c.slice/crio-c79c69711241e419533efce9996d187979188446228a625c0ac0b552dfabd688 WatchSource:0}: Error finding container c79c69711241e419533efce9996d187979188446228a625c0ac0b552dfabd688: Status 404 returned error can't find the container with id c79c69711241e419533efce9996d187979188446228a625c0ac0b552dfabd688 Feb 27 16:52:01 crc kubenswrapper[4814]: I0227 16:52:01.013757 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536852-8vpsm"] Feb 27 16:52:01 crc kubenswrapper[4814]: I0227 16:52:01.900766 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536852-8vpsm" event={"ID":"93b95ff2-824a-43ff-9110-ef73f1fcc34c","Type":"ContainerStarted","Data":"c79c69711241e419533efce9996d187979188446228a625c0ac0b552dfabd688"} Feb 27 16:52:02 crc kubenswrapper[4814]: I0227 16:52:02.916670 4814 generic.go:334] "Generic (PLEG): container finished" podID="93b95ff2-824a-43ff-9110-ef73f1fcc34c" containerID="194d3a56e8652d3331106c1113dd11dda49c95bc664e6972c88716ae024dc4e7" exitCode=0 Feb 27 16:52:02 crc kubenswrapper[4814]: I0227 16:52:02.916771 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536852-8vpsm" event={"ID":"93b95ff2-824a-43ff-9110-ef73f1fcc34c","Type":"ContainerDied","Data":"194d3a56e8652d3331106c1113dd11dda49c95bc664e6972c88716ae024dc4e7"} Feb 27 16:52:04 crc kubenswrapper[4814]: I0227 16:52:04.318999 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536852-8vpsm" Feb 27 16:52:04 crc kubenswrapper[4814]: I0227 16:52:04.452168 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xbc5\" (UniqueName: \"kubernetes.io/projected/93b95ff2-824a-43ff-9110-ef73f1fcc34c-kube-api-access-9xbc5\") pod \"93b95ff2-824a-43ff-9110-ef73f1fcc34c\" (UID: \"93b95ff2-824a-43ff-9110-ef73f1fcc34c\") " Feb 27 16:52:04 crc kubenswrapper[4814]: I0227 16:52:04.458718 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93b95ff2-824a-43ff-9110-ef73f1fcc34c-kube-api-access-9xbc5" (OuterVolumeSpecName: "kube-api-access-9xbc5") pod "93b95ff2-824a-43ff-9110-ef73f1fcc34c" (UID: "93b95ff2-824a-43ff-9110-ef73f1fcc34c"). InnerVolumeSpecName "kube-api-access-9xbc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:52:04 crc kubenswrapper[4814]: I0227 16:52:04.554646 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xbc5\" (UniqueName: \"kubernetes.io/projected/93b95ff2-824a-43ff-9110-ef73f1fcc34c-kube-api-access-9xbc5\") on node \"crc\" DevicePath \"\"" Feb 27 16:52:04 crc kubenswrapper[4814]: I0227 16:52:04.940580 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536852-8vpsm" event={"ID":"93b95ff2-824a-43ff-9110-ef73f1fcc34c","Type":"ContainerDied","Data":"c79c69711241e419533efce9996d187979188446228a625c0ac0b552dfabd688"} Feb 27 16:52:04 crc kubenswrapper[4814]: I0227 16:52:04.940635 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c79c69711241e419533efce9996d187979188446228a625c0ac0b552dfabd688" Feb 27 16:52:04 crc kubenswrapper[4814]: I0227 16:52:04.940661 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536852-8vpsm" Feb 27 16:52:05 crc kubenswrapper[4814]: I0227 16:52:05.418704 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536846-k7bdk"] Feb 27 16:52:05 crc kubenswrapper[4814]: I0227 16:52:05.435855 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536846-k7bdk"] Feb 27 16:52:06 crc kubenswrapper[4814]: I0227 16:52:06.508040 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb98ddfc-a6c6-4407-9e0e-8f339bed0f03" path="/var/lib/kubelet/pods/eb98ddfc-a6c6-4407-9e0e-8f339bed0f03/volumes" Feb 27 16:52:22 crc kubenswrapper[4814]: I0227 16:52:22.903285 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:52:22 crc kubenswrapper[4814]: I0227 16:52:22.904120 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:52:35 crc kubenswrapper[4814]: I0227 16:52:35.055234 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-9b47-account-create-update-8bqfg"] Feb 27 16:52:35 crc kubenswrapper[4814]: I0227 16:52:35.065574 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-9b47-account-create-update-8bqfg"] Feb 27 16:52:36 crc kubenswrapper[4814]: I0227 16:52:36.052310 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-9481-account-create-update-rzblc"] Feb 27 16:52:36 crc kubenswrapper[4814]: I0227 16:52:36.068289 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-f22wg"] Feb 27 16:52:36 crc kubenswrapper[4814]: I0227 16:52:36.079038 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-j5g9f"] Feb 27 16:52:36 crc kubenswrapper[4814]: I0227 16:52:36.091135 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-8tfhv"] Feb 27 16:52:36 crc kubenswrapper[4814]: I0227 16:52:36.098504 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-9481-account-create-update-rzblc"] Feb 27 16:52:36 crc kubenswrapper[4814]: I0227 16:52:36.104696 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-f22wg"] Feb 27 16:52:36 crc kubenswrapper[4814]: I0227 16:52:36.111426 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-8tfhv"] Feb 27 16:52:36 crc kubenswrapper[4814]: I0227 16:52:36.117759 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-1631-account-create-update-mpg4t"] Feb 27 16:52:36 crc kubenswrapper[4814]: I0227 16:52:36.125540 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-1631-account-create-update-mpg4t"] Feb 27 16:52:36 crc kubenswrapper[4814]: I0227 16:52:36.130746 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-j5g9f"] Feb 27 16:52:36 crc kubenswrapper[4814]: I0227 16:52:36.504954 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c835cab-f5ff-497c-a4c3-865e2a30f48f" path="/var/lib/kubelet/pods/0c835cab-f5ff-497c-a4c3-865e2a30f48f/volumes" Feb 27 16:52:36 crc kubenswrapper[4814]: I0227 16:52:36.506342 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="653257dd-39a8-4062-b70f-33d78ef04baf" path="/var/lib/kubelet/pods/653257dd-39a8-4062-b70f-33d78ef04baf/volumes" Feb 27 16:52:36 crc kubenswrapper[4814]: I0227 16:52:36.507219 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a048b31-2392-4739-8350-93136b0820b3" path="/var/lib/kubelet/pods/7a048b31-2392-4739-8350-93136b0820b3/volumes" Feb 27 16:52:36 crc kubenswrapper[4814]: I0227 16:52:36.508056 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dc075b4-dac3-46c3-9bad-9ffd34052643" path="/var/lib/kubelet/pods/9dc075b4-dac3-46c3-9bad-9ffd34052643/volumes" Feb 27 16:52:36 crc kubenswrapper[4814]: I0227 16:52:36.509597 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e" path="/var/lib/kubelet/pods/dfc40f4f-7c6b-4bdf-aa9e-c5e19e0cc19e/volumes" Feb 27 16:52:36 crc kubenswrapper[4814]: I0227 16:52:36.510513 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4f446f4-2834-4ce5-8010-9e4308087ea1" path="/var/lib/kubelet/pods/e4f446f4-2834-4ce5-8010-9e4308087ea1/volumes" Feb 27 16:52:41 crc kubenswrapper[4814]: I0227 16:52:41.035774 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-ln7b6"] Feb 27 16:52:41 crc kubenswrapper[4814]: I0227 16:52:41.045417 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-ln7b6"] Feb 27 16:52:42 crc kubenswrapper[4814]: I0227 16:52:42.535084 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaef6a7d-1318-40a2-a11a-2e9e1241f828" path="/var/lib/kubelet/pods/eaef6a7d-1318-40a2-a11a-2e9e1241f828/volumes" Feb 27 16:52:52 crc kubenswrapper[4814]: I0227 16:52:52.903018 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 16:52:52 crc kubenswrapper[4814]: I0227 16:52:52.903822 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 16:52:52 crc kubenswrapper[4814]: I0227 16:52:52.903911 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 16:52:52 crc kubenswrapper[4814]: I0227 16:52:52.905464 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08"} pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 16:52:52 crc kubenswrapper[4814]: I0227 16:52:52.905837 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" containerID="cri-o://aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" gracePeriod=600 Feb 27 16:52:53 crc kubenswrapper[4814]: E0227 16:52:53.040870 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:52:53 crc kubenswrapper[4814]: I0227 16:52:53.552179 4814 generic.go:334] "Generic (PLEG): container finished" podID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" exitCode=0 Feb 27 16:52:53 crc kubenswrapper[4814]: I0227 16:52:53.552239 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerDied","Data":"aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08"} Feb 27 16:52:53 crc kubenswrapper[4814]: I0227 16:52:53.552540 4814 scope.go:117] "RemoveContainer" containerID="e5df7c66a9ca17c8ac1eb1c76463afbe46a665cbf347524e84122f10413b151c" Feb 27 16:52:53 crc kubenswrapper[4814]: I0227 16:52:53.553183 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:52:53 crc kubenswrapper[4814]: E0227 16:52:53.553513 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:52:54 crc kubenswrapper[4814]: I0227 16:52:54.279771 4814 scope.go:117] "RemoveContainer" containerID="4a83fe3e3495cfc2556109aa00214204acd10f9e608f2528294960ea853e9c93" Feb 27 16:52:54 crc kubenswrapper[4814]: I0227 16:52:54.317764 4814 scope.go:117] "RemoveContainer" containerID="51a1628bee460ac25f28940c3a4aa3bf5a67d97208f32aadf93d4d78850aa29b" Feb 27 16:52:54 crc kubenswrapper[4814]: I0227 16:52:54.396585 4814 scope.go:117] "RemoveContainer" containerID="abbee1a7af813844aae45f33cfda21df439528d9ef07c4277f524d0c02fca5cd" Feb 27 16:52:54 crc kubenswrapper[4814]: I0227 16:52:54.446844 4814 scope.go:117] "RemoveContainer" containerID="87966d0a1f0b1cd11e6e08bb7c206f16b3e7acd459379f5a938175794bbb9f11" Feb 27 16:52:54 crc kubenswrapper[4814]: I0227 16:52:54.505644 4814 scope.go:117] "RemoveContainer" containerID="c48c2203662b83108b76b2b3cdb1e68ae49384a83a7d8c64850261a730f4ca1d" Feb 27 16:52:54 crc kubenswrapper[4814]: I0227 16:52:54.541505 4814 scope.go:117] "RemoveContainer" containerID="5979d3004d6574e45657308442d2ca78c23c039809420092a8a73557646b8791" Feb 27 16:52:54 crc kubenswrapper[4814]: I0227 16:52:54.617516 4814 scope.go:117] "RemoveContainer" containerID="f8f054552a5c11789a93dc8da2f52e6207e11331d4008eb36ea8a980b12ce547" Feb 27 16:52:54 crc kubenswrapper[4814]: I0227 16:52:54.638951 4814 scope.go:117] "RemoveContainer" containerID="969868d9d157b53533159f7030e94a33751fb3bba169453cdd525ac1bdf44476" Feb 27 16:53:04 crc kubenswrapper[4814]: I0227 16:53:04.487283 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:53:04 crc kubenswrapper[4814]: E0227 16:53:04.488076 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:53:05 crc kubenswrapper[4814]: I0227 16:53:05.067845 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-k85vk"] Feb 27 16:53:05 crc kubenswrapper[4814]: I0227 16:53:05.089369 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-k85vk"] Feb 27 16:53:06 crc kubenswrapper[4814]: I0227 16:53:06.512891 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9" path="/var/lib/kubelet/pods/7a3de8b8-bfb6-4e36-9e86-1b1851c50eb9/volumes" Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.030376 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-2e37-account-create-update-dxdg7"] Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.040852 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-tvmcg"] Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.050680 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-vphqv"] Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.065459 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-2845-account-create-update-pzlvv"] Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.076906 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-6479-account-create-update-5jbnw"] Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.088399 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-2e37-account-create-update-dxdg7"] Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.095052 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-vphqv"] Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.103889 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-lgxt2"] Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.112805 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-tvmcg"] Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.121359 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-2845-account-create-update-pzlvv"] Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.128761 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-6479-account-create-update-5jbnw"] Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.136271 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-lgxt2"] Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.500529 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9" path="/var/lib/kubelet/pods/0c7c3f6a-ab11-4b2b-b8c9-1d6ca23631e9/volumes" Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.501321 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f" path="/var/lib/kubelet/pods/0d7d2f97-f0db-4a91-a3c5-02b8b3ded11f/volumes" Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.501917 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d3720d6-e79c-44cc-8780-415f2b139a9f" path="/var/lib/kubelet/pods/8d3720d6-e79c-44cc-8780-415f2b139a9f/volumes" Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.502550 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c2ab590-4b44-4d31-9324-2abffd328742" path="/var/lib/kubelet/pods/9c2ab590-4b44-4d31-9324-2abffd328742/volumes" Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.503667 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b885ee2d-5c14-40b1-83c5-068d2bb0ef96" path="/var/lib/kubelet/pods/b885ee2d-5c14-40b1-83c5-068d2bb0ef96/volumes" Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.504270 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd1f40c7-a082-4356-848b-578cb58f1015" path="/var/lib/kubelet/pods/bd1f40c7-a082-4356-848b-578cb58f1015/volumes" Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.799675 4814 generic.go:334] "Generic (PLEG): container finished" podID="daf9ab93-de26-4929-a827-925d1650af5a" containerID="ac1eaf9f77f22aaa93155580aa1c7879f4e0d7724be68589bb5830957e266a8f" exitCode=0 Feb 27 16:53:12 crc kubenswrapper[4814]: I0227 16:53:12.799795 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" event={"ID":"daf9ab93-de26-4929-a827-925d1650af5a","Type":"ContainerDied","Data":"ac1eaf9f77f22aaa93155580aa1c7879f4e0d7724be68589bb5830957e266a8f"} Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.196179 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.372677 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/daf9ab93-de26-4929-a827-925d1650af5a-inventory\") pod \"daf9ab93-de26-4929-a827-925d1650af5a\" (UID: \"daf9ab93-de26-4929-a827-925d1650af5a\") " Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.372771 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q22z\" (UniqueName: \"kubernetes.io/projected/daf9ab93-de26-4929-a827-925d1650af5a-kube-api-access-7q22z\") pod \"daf9ab93-de26-4929-a827-925d1650af5a\" (UID: \"daf9ab93-de26-4929-a827-925d1650af5a\") " Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.372850 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/daf9ab93-de26-4929-a827-925d1650af5a-ssh-key-openstack-edpm-ipam\") pod \"daf9ab93-de26-4929-a827-925d1650af5a\" (UID: \"daf9ab93-de26-4929-a827-925d1650af5a\") " Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.379821 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daf9ab93-de26-4929-a827-925d1650af5a-kube-api-access-7q22z" (OuterVolumeSpecName: "kube-api-access-7q22z") pod "daf9ab93-de26-4929-a827-925d1650af5a" (UID: "daf9ab93-de26-4929-a827-925d1650af5a"). InnerVolumeSpecName "kube-api-access-7q22z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.404567 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf9ab93-de26-4929-a827-925d1650af5a-inventory" (OuterVolumeSpecName: "inventory") pod "daf9ab93-de26-4929-a827-925d1650af5a" (UID: "daf9ab93-de26-4929-a827-925d1650af5a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.408361 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf9ab93-de26-4929-a827-925d1650af5a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "daf9ab93-de26-4929-a827-925d1650af5a" (UID: "daf9ab93-de26-4929-a827-925d1650af5a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.475052 4814 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/daf9ab93-de26-4929-a827-925d1650af5a-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.475081 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q22z\" (UniqueName: \"kubernetes.io/projected/daf9ab93-de26-4929-a827-925d1650af5a-kube-api-access-7q22z\") on node \"crc\" DevicePath \"\"" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.475094 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/daf9ab93-de26-4929-a827-925d1650af5a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.820391 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" event={"ID":"daf9ab93-de26-4929-a827-925d1650af5a","Type":"ContainerDied","Data":"7d5080a90ba10a4e5b8b6af1710068b374ad6588da17fa89a73a2a013cc46815"} Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.820448 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d5080a90ba10a4e5b8b6af1710068b374ad6588da17fa89a73a2a013cc46815" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.820527 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.917767 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw"] Feb 27 16:53:14 crc kubenswrapper[4814]: E0227 16:53:14.918614 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93b95ff2-824a-43ff-9110-ef73f1fcc34c" containerName="oc" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.918645 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="93b95ff2-824a-43ff-9110-ef73f1fcc34c" containerName="oc" Feb 27 16:53:14 crc kubenswrapper[4814]: E0227 16:53:14.918684 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daf9ab93-de26-4929-a827-925d1650af5a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.918697 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="daf9ab93-de26-4929-a827-925d1650af5a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.918978 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="93b95ff2-824a-43ff-9110-ef73f1fcc34c" containerName="oc" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.919033 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="daf9ab93-de26-4929-a827-925d1650af5a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.919895 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.922248 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.922279 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.922304 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.923213 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 16:53:14 crc kubenswrapper[4814]: I0227 16:53:14.930614 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw"] Feb 27 16:53:15 crc kubenswrapper[4814]: I0227 16:53:15.085266 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c0fc8c4-5559-4dd9-9220-b47c3b537749-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw\" (UID: \"1c0fc8c4-5559-4dd9-9220-b47c3b537749\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" Feb 27 16:53:15 crc kubenswrapper[4814]: I0227 16:53:15.085328 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqp7c\" (UniqueName: \"kubernetes.io/projected/1c0fc8c4-5559-4dd9-9220-b47c3b537749-kube-api-access-mqp7c\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw\" (UID: \"1c0fc8c4-5559-4dd9-9220-b47c3b537749\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" Feb 27 16:53:15 crc kubenswrapper[4814]: I0227 16:53:15.085426 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1c0fc8c4-5559-4dd9-9220-b47c3b537749-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw\" (UID: \"1c0fc8c4-5559-4dd9-9220-b47c3b537749\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" Feb 27 16:53:15 crc kubenswrapper[4814]: I0227 16:53:15.187363 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1c0fc8c4-5559-4dd9-9220-b47c3b537749-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw\" (UID: \"1c0fc8c4-5559-4dd9-9220-b47c3b537749\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" Feb 27 16:53:15 crc kubenswrapper[4814]: I0227 16:53:15.187550 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c0fc8c4-5559-4dd9-9220-b47c3b537749-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw\" (UID: \"1c0fc8c4-5559-4dd9-9220-b47c3b537749\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" Feb 27 16:53:15 crc kubenswrapper[4814]: I0227 16:53:15.187597 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqp7c\" (UniqueName: \"kubernetes.io/projected/1c0fc8c4-5559-4dd9-9220-b47c3b537749-kube-api-access-mqp7c\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw\" (UID: \"1c0fc8c4-5559-4dd9-9220-b47c3b537749\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" Feb 27 16:53:15 crc kubenswrapper[4814]: I0227 16:53:15.192780 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1c0fc8c4-5559-4dd9-9220-b47c3b537749-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw\" (UID: \"1c0fc8c4-5559-4dd9-9220-b47c3b537749\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" Feb 27 16:53:15 crc kubenswrapper[4814]: I0227 16:53:15.193815 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c0fc8c4-5559-4dd9-9220-b47c3b537749-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw\" (UID: \"1c0fc8c4-5559-4dd9-9220-b47c3b537749\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" Feb 27 16:53:15 crc kubenswrapper[4814]: I0227 16:53:15.220887 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqp7c\" (UniqueName: \"kubernetes.io/projected/1c0fc8c4-5559-4dd9-9220-b47c3b537749-kube-api-access-mqp7c\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw\" (UID: \"1c0fc8c4-5559-4dd9-9220-b47c3b537749\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" Feb 27 16:53:15 crc kubenswrapper[4814]: I0227 16:53:15.237714 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" Feb 27 16:53:15 crc kubenswrapper[4814]: I0227 16:53:15.843007 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw"] Feb 27 16:53:15 crc kubenswrapper[4814]: I0227 16:53:15.856920 4814 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 16:53:16 crc kubenswrapper[4814]: I0227 16:53:16.840943 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" event={"ID":"1c0fc8c4-5559-4dd9-9220-b47c3b537749","Type":"ContainerStarted","Data":"222b57148d9010f877e6eb0dc97ce1e8ea290fa04737c1165f42eeffa47fe8b8"} Feb 27 16:53:16 crc kubenswrapper[4814]: I0227 16:53:16.841291 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" event={"ID":"1c0fc8c4-5559-4dd9-9220-b47c3b537749","Type":"ContainerStarted","Data":"e3b80aaab7867845af877fd0e475ed467117f0552b14c065be99e5bffcac4c7f"} Feb 27 16:53:16 crc kubenswrapper[4814]: I0227 16:53:16.859465 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" podStartSLOduration=2.450654868 podStartE2EDuration="2.859444123s" podCreationTimestamp="2026-02-27 16:53:14 +0000 UTC" firstStartedPulling="2026-02-27 16:53:15.856517539 +0000 UTC m=+1808.309142389" lastFinishedPulling="2026-02-27 16:53:16.265306814 +0000 UTC m=+1808.717931644" observedRunningTime="2026-02-27 16:53:16.85865985 +0000 UTC m=+1809.311284680" watchObservedRunningTime="2026-02-27 16:53:16.859444123 +0000 UTC m=+1809.312068953" Feb 27 16:53:17 crc kubenswrapper[4814]: I0227 16:53:17.488039 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:53:17 crc kubenswrapper[4814]: E0227 16:53:17.488425 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:53:20 crc kubenswrapper[4814]: I0227 16:53:20.049925 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-l4sht"] Feb 27 16:53:20 crc kubenswrapper[4814]: I0227 16:53:20.065453 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-l4sht"] Feb 27 16:53:20 crc kubenswrapper[4814]: I0227 16:53:20.505521 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6779899f-08d7-4de4-884c-070af4597386" path="/var/lib/kubelet/pods/6779899f-08d7-4de4-884c-070af4597386/volumes" Feb 27 16:53:28 crc kubenswrapper[4814]: I0227 16:53:28.498620 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:53:28 crc kubenswrapper[4814]: E0227 16:53:28.502239 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:53:41 crc kubenswrapper[4814]: I0227 16:53:41.487645 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:53:41 crc kubenswrapper[4814]: E0227 16:53:41.488982 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:53:53 crc kubenswrapper[4814]: I0227 16:53:53.488369 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:53:53 crc kubenswrapper[4814]: E0227 16:53:53.489472 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:53:54 crc kubenswrapper[4814]: I0227 16:53:54.826446 4814 scope.go:117] "RemoveContainer" containerID="7d9f4a94e9218f9a4b247299d02a5ad4403f74bc50a6ef5609679e4d8d88bf3a" Feb 27 16:53:54 crc kubenswrapper[4814]: I0227 16:53:54.854099 4814 scope.go:117] "RemoveContainer" containerID="d0ae7868c8f87fcbb95cab422c02698a0c3eac93473dcbe8ef0196aa8e9743eb" Feb 27 16:53:54 crc kubenswrapper[4814]: I0227 16:53:54.940616 4814 scope.go:117] "RemoveContainer" containerID="07f90fb740624019f8172a6b68f873b900dcfd78787dfc6961f3576d32e43e6d" Feb 27 16:53:54 crc kubenswrapper[4814]: I0227 16:53:54.989874 4814 scope.go:117] "RemoveContainer" containerID="002ed9a3bed92d76b420850d31727c7776c5ef76ad9fe80068395f30a52215b5" Feb 27 16:53:55 crc kubenswrapper[4814]: I0227 16:53:55.049380 4814 scope.go:117] "RemoveContainer" containerID="db2b3be43822ed9bd4853f6086c1e27e0ca957584c3e2042429501bca8827514" Feb 27 16:53:55 crc kubenswrapper[4814]: I0227 16:53:55.052032 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-th62b"] Feb 27 16:53:55 crc kubenswrapper[4814]: I0227 16:53:55.065514 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-th62b"] Feb 27 16:53:55 crc kubenswrapper[4814]: I0227 16:53:55.096568 4814 scope.go:117] "RemoveContainer" containerID="f3f6819b6255e864f982437f471dfed41ea027a8f03f66d18cea0054832d3c1d" Feb 27 16:53:55 crc kubenswrapper[4814]: I0227 16:53:55.138246 4814 scope.go:117] "RemoveContainer" containerID="76d4b032aba40b34faa39cbb6776ef5e48eedd9b1974b40a8516c0075c6b2d3c" Feb 27 16:53:55 crc kubenswrapper[4814]: I0227 16:53:55.190276 4814 scope.go:117] "RemoveContainer" containerID="6e34a5ba5d94a1d6d82632db4751abd111e8ee2716a87987714bd0ad494938e1" Feb 27 16:53:56 crc kubenswrapper[4814]: I0227 16:53:56.506683 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e413768-6d50-4725-a809-3658e913cfed" path="/var/lib/kubelet/pods/5e413768-6d50-4725-a809-3658e913cfed/volumes" Feb 27 16:53:58 crc kubenswrapper[4814]: I0227 16:53:58.045298 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-pm95c"] Feb 27 16:53:58 crc kubenswrapper[4814]: I0227 16:53:58.061085 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-pm95c"] Feb 27 16:53:58 crc kubenswrapper[4814]: I0227 16:53:58.507094 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c357e0a2-9571-4e9e-b48c-0f92e2c99afc" path="/var/lib/kubelet/pods/c357e0a2-9571-4e9e-b48c-0f92e2c99afc/volumes" Feb 27 16:54:00 crc kubenswrapper[4814]: I0227 16:54:00.163707 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536854-pkbjj"] Feb 27 16:54:00 crc kubenswrapper[4814]: I0227 16:54:00.166757 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536854-pkbjj" Feb 27 16:54:00 crc kubenswrapper[4814]: I0227 16:54:00.171027 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 16:54:00 crc kubenswrapper[4814]: I0227 16:54:00.171460 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:54:00 crc kubenswrapper[4814]: I0227 16:54:00.171480 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:54:00 crc kubenswrapper[4814]: I0227 16:54:00.180191 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536854-pkbjj"] Feb 27 16:54:00 crc kubenswrapper[4814]: I0227 16:54:00.258891 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjbkx\" (UniqueName: \"kubernetes.io/projected/3d93ea82-2104-4731-93ae-3e469a251396-kube-api-access-pjbkx\") pod \"auto-csr-approver-29536854-pkbjj\" (UID: \"3d93ea82-2104-4731-93ae-3e469a251396\") " pod="openshift-infra/auto-csr-approver-29536854-pkbjj" Feb 27 16:54:00 crc kubenswrapper[4814]: I0227 16:54:00.360993 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjbkx\" (UniqueName: \"kubernetes.io/projected/3d93ea82-2104-4731-93ae-3e469a251396-kube-api-access-pjbkx\") pod \"auto-csr-approver-29536854-pkbjj\" (UID: \"3d93ea82-2104-4731-93ae-3e469a251396\") " pod="openshift-infra/auto-csr-approver-29536854-pkbjj" Feb 27 16:54:00 crc kubenswrapper[4814]: I0227 16:54:00.402891 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjbkx\" (UniqueName: \"kubernetes.io/projected/3d93ea82-2104-4731-93ae-3e469a251396-kube-api-access-pjbkx\") pod \"auto-csr-approver-29536854-pkbjj\" (UID: \"3d93ea82-2104-4731-93ae-3e469a251396\") " pod="openshift-infra/auto-csr-approver-29536854-pkbjj" Feb 27 16:54:00 crc kubenswrapper[4814]: I0227 16:54:00.513196 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536854-pkbjj" Feb 27 16:54:00 crc kubenswrapper[4814]: I0227 16:54:00.942194 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536854-pkbjj"] Feb 27 16:54:01 crc kubenswrapper[4814]: I0227 16:54:01.392423 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536854-pkbjj" event={"ID":"3d93ea82-2104-4731-93ae-3e469a251396","Type":"ContainerStarted","Data":"1b2a23fd204ccf1fef3e42f5118b410f674742df3c55b4d003ccd674dc371336"} Feb 27 16:54:02 crc kubenswrapper[4814]: I0227 16:54:02.048169 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-88h8r"] Feb 27 16:54:02 crc kubenswrapper[4814]: I0227 16:54:02.057004 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-88h8r"] Feb 27 16:54:02 crc kubenswrapper[4814]: I0227 16:54:02.426721 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536854-pkbjj" event={"ID":"3d93ea82-2104-4731-93ae-3e469a251396","Type":"ContainerStarted","Data":"1b66aa8513771120b04a51a53cb2e88cc6798d2624ffbbdcc62fc18f82ad397e"} Feb 27 16:54:02 crc kubenswrapper[4814]: I0227 16:54:02.442943 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536854-pkbjj" podStartSLOduration=1.319443594 podStartE2EDuration="2.442920024s" podCreationTimestamp="2026-02-27 16:54:00 +0000 UTC" firstStartedPulling="2026-02-27 16:54:00.95520604 +0000 UTC m=+1853.407830890" lastFinishedPulling="2026-02-27 16:54:02.07868249 +0000 UTC m=+1854.531307320" observedRunningTime="2026-02-27 16:54:02.43786682 +0000 UTC m=+1854.890491660" watchObservedRunningTime="2026-02-27 16:54:02.442920024 +0000 UTC m=+1854.895544854" Feb 27 16:54:02 crc kubenswrapper[4814]: I0227 16:54:02.507732 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f56fd06-b678-48f2-9f0d-0234a53b528b" path="/var/lib/kubelet/pods/6f56fd06-b678-48f2-9f0d-0234a53b528b/volumes" Feb 27 16:54:03 crc kubenswrapper[4814]: I0227 16:54:03.443240 4814 generic.go:334] "Generic (PLEG): container finished" podID="3d93ea82-2104-4731-93ae-3e469a251396" containerID="1b66aa8513771120b04a51a53cb2e88cc6798d2624ffbbdcc62fc18f82ad397e" exitCode=0 Feb 27 16:54:03 crc kubenswrapper[4814]: I0227 16:54:03.443322 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536854-pkbjj" event={"ID":"3d93ea82-2104-4731-93ae-3e469a251396","Type":"ContainerDied","Data":"1b66aa8513771120b04a51a53cb2e88cc6798d2624ffbbdcc62fc18f82ad397e"} Feb 27 16:54:04 crc kubenswrapper[4814]: I0227 16:54:04.872477 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536854-pkbjj" Feb 27 16:54:04 crc kubenswrapper[4814]: I0227 16:54:04.957205 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjbkx\" (UniqueName: \"kubernetes.io/projected/3d93ea82-2104-4731-93ae-3e469a251396-kube-api-access-pjbkx\") pod \"3d93ea82-2104-4731-93ae-3e469a251396\" (UID: \"3d93ea82-2104-4731-93ae-3e469a251396\") " Feb 27 16:54:04 crc kubenswrapper[4814]: I0227 16:54:04.965404 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d93ea82-2104-4731-93ae-3e469a251396-kube-api-access-pjbkx" (OuterVolumeSpecName: "kube-api-access-pjbkx") pod "3d93ea82-2104-4731-93ae-3e469a251396" (UID: "3d93ea82-2104-4731-93ae-3e469a251396"). InnerVolumeSpecName "kube-api-access-pjbkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:54:05 crc kubenswrapper[4814]: I0227 16:54:05.059994 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjbkx\" (UniqueName: \"kubernetes.io/projected/3d93ea82-2104-4731-93ae-3e469a251396-kube-api-access-pjbkx\") on node \"crc\" DevicePath \"\"" Feb 27 16:54:05 crc kubenswrapper[4814]: I0227 16:54:05.468747 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536854-pkbjj" event={"ID":"3d93ea82-2104-4731-93ae-3e469a251396","Type":"ContainerDied","Data":"1b2a23fd204ccf1fef3e42f5118b410f674742df3c55b4d003ccd674dc371336"} Feb 27 16:54:05 crc kubenswrapper[4814]: I0227 16:54:05.469085 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b2a23fd204ccf1fef3e42f5118b410f674742df3c55b4d003ccd674dc371336" Feb 27 16:54:05 crc kubenswrapper[4814]: I0227 16:54:05.468827 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536854-pkbjj" Feb 27 16:54:05 crc kubenswrapper[4814]: I0227 16:54:05.508220 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536848-wt9mj"] Feb 27 16:54:05 crc kubenswrapper[4814]: I0227 16:54:05.515958 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536848-wt9mj"] Feb 27 16:54:06 crc kubenswrapper[4814]: I0227 16:54:06.489221 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:54:06 crc kubenswrapper[4814]: E0227 16:54:06.490041 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:54:06 crc kubenswrapper[4814]: I0227 16:54:06.510574 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9036f38-d8ae-410f-a826-3bd8599e46ea" path="/var/lib/kubelet/pods/e9036f38-d8ae-410f-a826-3bd8599e46ea/volumes" Feb 27 16:54:09 crc kubenswrapper[4814]: I0227 16:54:09.058578 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-pq8r4"] Feb 27 16:54:09 crc kubenswrapper[4814]: I0227 16:54:09.074247 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-pq8r4"] Feb 27 16:54:10 crc kubenswrapper[4814]: I0227 16:54:10.511077 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f5812c2-9952-4125-b055-dd2888fad3f5" path="/var/lib/kubelet/pods/6f5812c2-9952-4125-b055-dd2888fad3f5/volumes" Feb 27 16:54:19 crc kubenswrapper[4814]: I0227 16:54:19.487466 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:54:19 crc kubenswrapper[4814]: E0227 16:54:19.488151 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:54:22 crc kubenswrapper[4814]: I0227 16:54:22.038666 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-vqkl5"] Feb 27 16:54:22 crc kubenswrapper[4814]: I0227 16:54:22.046929 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-vqkl5"] Feb 27 16:54:22 crc kubenswrapper[4814]: I0227 16:54:22.506443 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bcf9bc6-1fb4-47b4-b12d-017902e4992e" path="/var/lib/kubelet/pods/7bcf9bc6-1fb4-47b4-b12d-017902e4992e/volumes" Feb 27 16:54:27 crc kubenswrapper[4814]: I0227 16:54:27.778536 4814 generic.go:334] "Generic (PLEG): container finished" podID="1c0fc8c4-5559-4dd9-9220-b47c3b537749" containerID="222b57148d9010f877e6eb0dc97ce1e8ea290fa04737c1165f42eeffa47fe8b8" exitCode=0 Feb 27 16:54:27 crc kubenswrapper[4814]: I0227 16:54:27.778647 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" event={"ID":"1c0fc8c4-5559-4dd9-9220-b47c3b537749","Type":"ContainerDied","Data":"222b57148d9010f877e6eb0dc97ce1e8ea290fa04737c1165f42eeffa47fe8b8"} Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.299754 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.354294 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqp7c\" (UniqueName: \"kubernetes.io/projected/1c0fc8c4-5559-4dd9-9220-b47c3b537749-kube-api-access-mqp7c\") pod \"1c0fc8c4-5559-4dd9-9220-b47c3b537749\" (UID: \"1c0fc8c4-5559-4dd9-9220-b47c3b537749\") " Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.354364 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c0fc8c4-5559-4dd9-9220-b47c3b537749-inventory\") pod \"1c0fc8c4-5559-4dd9-9220-b47c3b537749\" (UID: \"1c0fc8c4-5559-4dd9-9220-b47c3b537749\") " Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.354395 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1c0fc8c4-5559-4dd9-9220-b47c3b537749-ssh-key-openstack-edpm-ipam\") pod \"1c0fc8c4-5559-4dd9-9220-b47c3b537749\" (UID: \"1c0fc8c4-5559-4dd9-9220-b47c3b537749\") " Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.378677 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c0fc8c4-5559-4dd9-9220-b47c3b537749-kube-api-access-mqp7c" (OuterVolumeSpecName: "kube-api-access-mqp7c") pod "1c0fc8c4-5559-4dd9-9220-b47c3b537749" (UID: "1c0fc8c4-5559-4dd9-9220-b47c3b537749"). InnerVolumeSpecName "kube-api-access-mqp7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.394835 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c0fc8c4-5559-4dd9-9220-b47c3b537749-inventory" (OuterVolumeSpecName: "inventory") pod "1c0fc8c4-5559-4dd9-9220-b47c3b537749" (UID: "1c0fc8c4-5559-4dd9-9220-b47c3b537749"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.420371 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c0fc8c4-5559-4dd9-9220-b47c3b537749-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "1c0fc8c4-5559-4dd9-9220-b47c3b537749" (UID: "1c0fc8c4-5559-4dd9-9220-b47c3b537749"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.456516 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqp7c\" (UniqueName: \"kubernetes.io/projected/1c0fc8c4-5559-4dd9-9220-b47c3b537749-kube-api-access-mqp7c\") on node \"crc\" DevicePath \"\"" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.456557 4814 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c0fc8c4-5559-4dd9-9220-b47c3b537749-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.456571 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1c0fc8c4-5559-4dd9-9220-b47c3b537749-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.802219 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" event={"ID":"1c0fc8c4-5559-4dd9-9220-b47c3b537749","Type":"ContainerDied","Data":"e3b80aaab7867845af877fd0e475ed467117f0552b14c065be99e5bffcac4c7f"} Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.802276 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3b80aaab7867845af877fd0e475ed467117f0552b14c065be99e5bffcac4c7f" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.802336 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.918028 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr"] Feb 27 16:54:29 crc kubenswrapper[4814]: E0227 16:54:29.918873 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c0fc8c4-5559-4dd9-9220-b47c3b537749" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.918907 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c0fc8c4-5559-4dd9-9220-b47c3b537749" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 27 16:54:29 crc kubenswrapper[4814]: E0227 16:54:29.918955 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d93ea82-2104-4731-93ae-3e469a251396" containerName="oc" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.918967 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d93ea82-2104-4731-93ae-3e469a251396" containerName="oc" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.919227 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d93ea82-2104-4731-93ae-3e469a251396" containerName="oc" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.919333 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c0fc8c4-5559-4dd9-9220-b47c3b537749" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.920478 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.923348 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.923580 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.924504 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.924692 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.926862 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr"] Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.964975 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phrgr\" (UID: \"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.965047 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pg5k\" (UniqueName: \"kubernetes.io/projected/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-kube-api-access-4pg5k\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phrgr\" (UID: \"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" Feb 27 16:54:29 crc kubenswrapper[4814]: I0227 16:54:29.965147 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phrgr\" (UID: \"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" Feb 27 16:54:30 crc kubenswrapper[4814]: I0227 16:54:30.066079 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pg5k\" (UniqueName: \"kubernetes.io/projected/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-kube-api-access-4pg5k\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phrgr\" (UID: \"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" Feb 27 16:54:30 crc kubenswrapper[4814]: I0227 16:54:30.066170 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phrgr\" (UID: \"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" Feb 27 16:54:30 crc kubenswrapper[4814]: I0227 16:54:30.066307 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phrgr\" (UID: \"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" Feb 27 16:54:30 crc kubenswrapper[4814]: I0227 16:54:30.070974 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phrgr\" (UID: \"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" Feb 27 16:54:30 crc kubenswrapper[4814]: I0227 16:54:30.071458 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phrgr\" (UID: \"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" Feb 27 16:54:30 crc kubenswrapper[4814]: I0227 16:54:30.097958 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pg5k\" (UniqueName: \"kubernetes.io/projected/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-kube-api-access-4pg5k\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phrgr\" (UID: \"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" Feb 27 16:54:30 crc kubenswrapper[4814]: I0227 16:54:30.242603 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" Feb 27 16:54:30 crc kubenswrapper[4814]: I0227 16:54:30.677713 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr"] Feb 27 16:54:30 crc kubenswrapper[4814]: W0227 16:54:30.681954 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4edcd988_23c8_4b8c_a3cd_a4d1e727ca3d.slice/crio-a6c3d4236a6566791e3039464bd2c3080afe6db91fad28d8be501246ed4b4274 WatchSource:0}: Error finding container a6c3d4236a6566791e3039464bd2c3080afe6db91fad28d8be501246ed4b4274: Status 404 returned error can't find the container with id a6c3d4236a6566791e3039464bd2c3080afe6db91fad28d8be501246ed4b4274 Feb 27 16:54:30 crc kubenswrapper[4814]: I0227 16:54:30.812905 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" event={"ID":"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d","Type":"ContainerStarted","Data":"a6c3d4236a6566791e3039464bd2c3080afe6db91fad28d8be501246ed4b4274"} Feb 27 16:54:31 crc kubenswrapper[4814]: I0227 16:54:31.826460 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" event={"ID":"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d","Type":"ContainerStarted","Data":"4310e8d3ac62ce838a2bcb33536de27ee5984162c3623d5e890594194e36b820"} Feb 27 16:54:31 crc kubenswrapper[4814]: I0227 16:54:31.854312 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" podStartSLOduration=2.245489978 podStartE2EDuration="2.854286376s" podCreationTimestamp="2026-02-27 16:54:29 +0000 UTC" firstStartedPulling="2026-02-27 16:54:30.686747569 +0000 UTC m=+1883.139372409" lastFinishedPulling="2026-02-27 16:54:31.295543967 +0000 UTC m=+1883.748168807" observedRunningTime="2026-02-27 16:54:31.843386403 +0000 UTC m=+1884.296011263" watchObservedRunningTime="2026-02-27 16:54:31.854286376 +0000 UTC m=+1884.306911216" Feb 27 16:54:34 crc kubenswrapper[4814]: I0227 16:54:34.492171 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:54:34 crc kubenswrapper[4814]: E0227 16:54:34.492901 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:54:36 crc kubenswrapper[4814]: I0227 16:54:36.889248 4814 generic.go:334] "Generic (PLEG): container finished" podID="4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d" containerID="4310e8d3ac62ce838a2bcb33536de27ee5984162c3623d5e890594194e36b820" exitCode=0 Feb 27 16:54:36 crc kubenswrapper[4814]: I0227 16:54:36.889290 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" event={"ID":"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d","Type":"ContainerDied","Data":"4310e8d3ac62ce838a2bcb33536de27ee5984162c3623d5e890594194e36b820"} Feb 27 16:54:38 crc kubenswrapper[4814]: I0227 16:54:38.366204 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" Feb 27 16:54:38 crc kubenswrapper[4814]: I0227 16:54:38.465868 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-inventory\") pod \"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d\" (UID: \"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d\") " Feb 27 16:54:38 crc kubenswrapper[4814]: I0227 16:54:38.466052 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pg5k\" (UniqueName: \"kubernetes.io/projected/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-kube-api-access-4pg5k\") pod \"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d\" (UID: \"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d\") " Feb 27 16:54:38 crc kubenswrapper[4814]: I0227 16:54:38.466157 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-ssh-key-openstack-edpm-ipam\") pod \"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d\" (UID: \"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d\") " Feb 27 16:54:38 crc kubenswrapper[4814]: I0227 16:54:38.483501 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-kube-api-access-4pg5k" (OuterVolumeSpecName: "kube-api-access-4pg5k") pod "4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d" (UID: "4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d"). InnerVolumeSpecName "kube-api-access-4pg5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:54:38 crc kubenswrapper[4814]: I0227 16:54:38.497832 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d" (UID: "4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:54:38 crc kubenswrapper[4814]: I0227 16:54:38.506086 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-inventory" (OuterVolumeSpecName: "inventory") pod "4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d" (UID: "4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:54:38 crc kubenswrapper[4814]: I0227 16:54:38.569141 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pg5k\" (UniqueName: \"kubernetes.io/projected/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-kube-api-access-4pg5k\") on node \"crc\" DevicePath \"\"" Feb 27 16:54:38 crc kubenswrapper[4814]: I0227 16:54:38.569204 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 16:54:38 crc kubenswrapper[4814]: I0227 16:54:38.569230 4814 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 16:54:38 crc kubenswrapper[4814]: I0227 16:54:38.914662 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" event={"ID":"4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d","Type":"ContainerDied","Data":"a6c3d4236a6566791e3039464bd2c3080afe6db91fad28d8be501246ed4b4274"} Feb 27 16:54:38 crc kubenswrapper[4814]: I0227 16:54:38.914715 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6c3d4236a6566791e3039464bd2c3080afe6db91fad28d8be501246ed4b4274" Feb 27 16:54:38 crc kubenswrapper[4814]: I0227 16:54:38.914782 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phrgr" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.027789 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt"] Feb 27 16:54:39 crc kubenswrapper[4814]: E0227 16:54:39.028156 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.028174 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.028360 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.028943 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.031749 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.036595 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.036625 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.036804 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.058512 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt"] Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.194764 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sqtqt\" (UID: \"f4d8ff3a-2b21-45ab-8332-5bedf355c09c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.194847 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmjln\" (UniqueName: \"kubernetes.io/projected/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-kube-api-access-nmjln\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sqtqt\" (UID: \"f4d8ff3a-2b21-45ab-8332-5bedf355c09c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.194873 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sqtqt\" (UID: \"f4d8ff3a-2b21-45ab-8332-5bedf355c09c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.296473 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmjln\" (UniqueName: \"kubernetes.io/projected/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-kube-api-access-nmjln\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sqtqt\" (UID: \"f4d8ff3a-2b21-45ab-8332-5bedf355c09c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.296544 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sqtqt\" (UID: \"f4d8ff3a-2b21-45ab-8332-5bedf355c09c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.297338 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sqtqt\" (UID: \"f4d8ff3a-2b21-45ab-8332-5bedf355c09c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.301878 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sqtqt\" (UID: \"f4d8ff3a-2b21-45ab-8332-5bedf355c09c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.303850 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sqtqt\" (UID: \"f4d8ff3a-2b21-45ab-8332-5bedf355c09c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.318922 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmjln\" (UniqueName: \"kubernetes.io/projected/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-kube-api-access-nmjln\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sqtqt\" (UID: \"f4d8ff3a-2b21-45ab-8332-5bedf355c09c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.344057 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" Feb 27 16:54:39 crc kubenswrapper[4814]: I0227 16:54:39.941152 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt"] Feb 27 16:54:40 crc kubenswrapper[4814]: I0227 16:54:40.941969 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" event={"ID":"f4d8ff3a-2b21-45ab-8332-5bedf355c09c","Type":"ContainerStarted","Data":"4262886ad9da8c01017c55ff662a675d638dc8e1935a9d8ecadcb39df7ec1446"} Feb 27 16:54:41 crc kubenswrapper[4814]: I0227 16:54:41.953714 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" event={"ID":"f4d8ff3a-2b21-45ab-8332-5bedf355c09c","Type":"ContainerStarted","Data":"4e09917a07d52e2ec81a2c1184e5a97f383bea284ad0b6cc93b886849477685c"} Feb 27 16:54:41 crc kubenswrapper[4814]: I0227 16:54:41.978758 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" podStartSLOduration=2.710129746 podStartE2EDuration="3.978732252s" podCreationTimestamp="2026-02-27 16:54:38 +0000 UTC" firstStartedPulling="2026-02-27 16:54:39.933722224 +0000 UTC m=+1892.386347064" lastFinishedPulling="2026-02-27 16:54:41.2023247 +0000 UTC m=+1893.654949570" observedRunningTime="2026-02-27 16:54:41.967822549 +0000 UTC m=+1894.420447429" watchObservedRunningTime="2026-02-27 16:54:41.978732252 +0000 UTC m=+1894.431357112" Feb 27 16:54:42 crc kubenswrapper[4814]: I0227 16:54:42.452068 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vzpq8"] Feb 27 16:54:42 crc kubenswrapper[4814]: I0227 16:54:42.455316 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:54:42 crc kubenswrapper[4814]: I0227 16:54:42.503672 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vzpq8"] Feb 27 16:54:42 crc kubenswrapper[4814]: I0227 16:54:42.565790 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a144b53-17a0-4edd-a9fb-bb83cab694a5-utilities\") pod \"community-operators-vzpq8\" (UID: \"7a144b53-17a0-4edd-a9fb-bb83cab694a5\") " pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:54:42 crc kubenswrapper[4814]: I0227 16:54:42.565840 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a144b53-17a0-4edd-a9fb-bb83cab694a5-catalog-content\") pod \"community-operators-vzpq8\" (UID: \"7a144b53-17a0-4edd-a9fb-bb83cab694a5\") " pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:54:42 crc kubenswrapper[4814]: I0227 16:54:42.565880 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb58l\" (UniqueName: \"kubernetes.io/projected/7a144b53-17a0-4edd-a9fb-bb83cab694a5-kube-api-access-wb58l\") pod \"community-operators-vzpq8\" (UID: \"7a144b53-17a0-4edd-a9fb-bb83cab694a5\") " pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:54:42 crc kubenswrapper[4814]: I0227 16:54:42.669061 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a144b53-17a0-4edd-a9fb-bb83cab694a5-utilities\") pod \"community-operators-vzpq8\" (UID: \"7a144b53-17a0-4edd-a9fb-bb83cab694a5\") " pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:54:42 crc kubenswrapper[4814]: I0227 16:54:42.669117 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a144b53-17a0-4edd-a9fb-bb83cab694a5-catalog-content\") pod \"community-operators-vzpq8\" (UID: \"7a144b53-17a0-4edd-a9fb-bb83cab694a5\") " pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:54:42 crc kubenswrapper[4814]: I0227 16:54:42.669148 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb58l\" (UniqueName: \"kubernetes.io/projected/7a144b53-17a0-4edd-a9fb-bb83cab694a5-kube-api-access-wb58l\") pod \"community-operators-vzpq8\" (UID: \"7a144b53-17a0-4edd-a9fb-bb83cab694a5\") " pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:54:42 crc kubenswrapper[4814]: I0227 16:54:42.669639 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a144b53-17a0-4edd-a9fb-bb83cab694a5-catalog-content\") pod \"community-operators-vzpq8\" (UID: \"7a144b53-17a0-4edd-a9fb-bb83cab694a5\") " pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:54:42 crc kubenswrapper[4814]: I0227 16:54:42.669741 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a144b53-17a0-4edd-a9fb-bb83cab694a5-utilities\") pod \"community-operators-vzpq8\" (UID: \"7a144b53-17a0-4edd-a9fb-bb83cab694a5\") " pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:54:42 crc kubenswrapper[4814]: I0227 16:54:42.702246 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb58l\" (UniqueName: \"kubernetes.io/projected/7a144b53-17a0-4edd-a9fb-bb83cab694a5-kube-api-access-wb58l\") pod \"community-operators-vzpq8\" (UID: \"7a144b53-17a0-4edd-a9fb-bb83cab694a5\") " pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:54:42 crc kubenswrapper[4814]: I0227 16:54:42.788354 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:54:43 crc kubenswrapper[4814]: I0227 16:54:43.303597 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vzpq8"] Feb 27 16:54:43 crc kubenswrapper[4814]: W0227 16:54:43.308349 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a144b53_17a0_4edd_a9fb_bb83cab694a5.slice/crio-7a9b591e6416a05dcc174e054e043ea0be660442624de4589099c5f24454c94a WatchSource:0}: Error finding container 7a9b591e6416a05dcc174e054e043ea0be660442624de4589099c5f24454c94a: Status 404 returned error can't find the container with id 7a9b591e6416a05dcc174e054e043ea0be660442624de4589099c5f24454c94a Feb 27 16:54:43 crc kubenswrapper[4814]: I0227 16:54:43.976865 4814 generic.go:334] "Generic (PLEG): container finished" podID="7a144b53-17a0-4edd-a9fb-bb83cab694a5" containerID="948b40794b5f8529e7c6dbd13cf3768c6edbd4655812250602b46ef447b813d9" exitCode=0 Feb 27 16:54:43 crc kubenswrapper[4814]: I0227 16:54:43.976919 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzpq8" event={"ID":"7a144b53-17a0-4edd-a9fb-bb83cab694a5","Type":"ContainerDied","Data":"948b40794b5f8529e7c6dbd13cf3768c6edbd4655812250602b46ef447b813d9"} Feb 27 16:54:43 crc kubenswrapper[4814]: I0227 16:54:43.976949 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzpq8" event={"ID":"7a144b53-17a0-4edd-a9fb-bb83cab694a5","Type":"ContainerStarted","Data":"7a9b591e6416a05dcc174e054e043ea0be660442624de4589099c5f24454c94a"} Feb 27 16:54:46 crc kubenswrapper[4814]: I0227 16:54:46.488207 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:54:46 crc kubenswrapper[4814]: E0227 16:54:46.489176 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:54:47 crc kubenswrapper[4814]: I0227 16:54:47.029808 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzpq8" event={"ID":"7a144b53-17a0-4edd-a9fb-bb83cab694a5","Type":"ContainerStarted","Data":"f9094016e0692db1d09824f955d031d5c4fbd67e11c8e567b466fcb444f6f8af"} Feb 27 16:54:48 crc kubenswrapper[4814]: I0227 16:54:48.045909 4814 generic.go:334] "Generic (PLEG): container finished" podID="7a144b53-17a0-4edd-a9fb-bb83cab694a5" containerID="f9094016e0692db1d09824f955d031d5c4fbd67e11c8e567b466fcb444f6f8af" exitCode=0 Feb 27 16:54:48 crc kubenswrapper[4814]: I0227 16:54:48.045957 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzpq8" event={"ID":"7a144b53-17a0-4edd-a9fb-bb83cab694a5","Type":"ContainerDied","Data":"f9094016e0692db1d09824f955d031d5c4fbd67e11c8e567b466fcb444f6f8af"} Feb 27 16:54:50 crc kubenswrapper[4814]: I0227 16:54:50.068897 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzpq8" event={"ID":"7a144b53-17a0-4edd-a9fb-bb83cab694a5","Type":"ContainerStarted","Data":"48c3add4be409a817573417a17f4373ae8894a2ca1ad763cca808b40ec9a4e75"} Feb 27 16:54:50 crc kubenswrapper[4814]: I0227 16:54:50.092790 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vzpq8" podStartSLOduration=3.166987645 podStartE2EDuration="8.092764387s" podCreationTimestamp="2026-02-27 16:54:42 +0000 UTC" firstStartedPulling="2026-02-27 16:54:43.978901749 +0000 UTC m=+1896.431526579" lastFinishedPulling="2026-02-27 16:54:48.904678481 +0000 UTC m=+1901.357303321" observedRunningTime="2026-02-27 16:54:50.086465914 +0000 UTC m=+1902.539090764" watchObservedRunningTime="2026-02-27 16:54:50.092764387 +0000 UTC m=+1902.545389227" Feb 27 16:54:52 crc kubenswrapper[4814]: I0227 16:54:52.789279 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:54:52 crc kubenswrapper[4814]: I0227 16:54:52.790409 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:54:52 crc kubenswrapper[4814]: I0227 16:54:52.943059 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:54:53 crc kubenswrapper[4814]: I0227 16:54:53.038881 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-bjcn9"] Feb 27 16:54:53 crc kubenswrapper[4814]: I0227 16:54:53.068715 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-lgx28"] Feb 27 16:54:53 crc kubenswrapper[4814]: I0227 16:54:53.077375 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-bjcn9"] Feb 27 16:54:53 crc kubenswrapper[4814]: I0227 16:54:53.083991 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-lgx28"] Feb 27 16:54:54 crc kubenswrapper[4814]: I0227 16:54:54.054942 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-b173-account-create-update-n5r76"] Feb 27 16:54:54 crc kubenswrapper[4814]: I0227 16:54:54.065927 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-bxqnd"] Feb 27 16:54:54 crc kubenswrapper[4814]: I0227 16:54:54.076673 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-5b77-account-create-update-t2vcb"] Feb 27 16:54:54 crc kubenswrapper[4814]: I0227 16:54:54.087711 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-5b77-account-create-update-t2vcb"] Feb 27 16:54:54 crc kubenswrapper[4814]: I0227 16:54:54.096173 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-bb22-account-create-update-xd4zd"] Feb 27 16:54:54 crc kubenswrapper[4814]: I0227 16:54:54.103149 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-b173-account-create-update-n5r76"] Feb 27 16:54:54 crc kubenswrapper[4814]: I0227 16:54:54.117078 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-bxqnd"] Feb 27 16:54:54 crc kubenswrapper[4814]: I0227 16:54:54.124025 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-bb22-account-create-update-xd4zd"] Feb 27 16:54:54 crc kubenswrapper[4814]: I0227 16:54:54.497999 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3977ccdd-8f4e-40c7-a0f8-7faa795620f0" path="/var/lib/kubelet/pods/3977ccdd-8f4e-40c7-a0f8-7faa795620f0/volumes" Feb 27 16:54:54 crc kubenswrapper[4814]: I0227 16:54:54.498761 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49d900c5-2698-4c77-8d0f-713d916ff26d" path="/var/lib/kubelet/pods/49d900c5-2698-4c77-8d0f-713d916ff26d/volumes" Feb 27 16:54:54 crc kubenswrapper[4814]: I0227 16:54:54.499460 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50b9ebf8-9e8b-476f-a868-688f3439aca4" path="/var/lib/kubelet/pods/50b9ebf8-9e8b-476f-a868-688f3439aca4/volumes" Feb 27 16:54:54 crc kubenswrapper[4814]: I0227 16:54:54.500249 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d287b36-3c56-48b2-aaff-8f8eaadb8f50" path="/var/lib/kubelet/pods/5d287b36-3c56-48b2-aaff-8f8eaadb8f50/volumes" Feb 27 16:54:54 crc kubenswrapper[4814]: I0227 16:54:54.501783 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80f1c20c-d371-4a0a-9c3f-e3002b086c1c" path="/var/lib/kubelet/pods/80f1c20c-d371-4a0a-9c3f-e3002b086c1c/volumes" Feb 27 16:54:54 crc kubenswrapper[4814]: I0227 16:54:54.503772 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcb4e437-b5fc-412b-9ef8-7dc917b08750" path="/var/lib/kubelet/pods/dcb4e437-b5fc-412b-9ef8-7dc917b08750/volumes" Feb 27 16:54:55 crc kubenswrapper[4814]: I0227 16:54:55.379012 4814 scope.go:117] "RemoveContainer" containerID="f8e24f09444ee738f777761fa58406e7797de5453d3ece79e9a37b9268433aec" Feb 27 16:54:55 crc kubenswrapper[4814]: I0227 16:54:55.426302 4814 scope.go:117] "RemoveContainer" containerID="67fab44cb9ddab92dbc952ba5825863ced0e115a1f4e9dac622240cd0f1e0176" Feb 27 16:54:55 crc kubenswrapper[4814]: I0227 16:54:55.515501 4814 scope.go:117] "RemoveContainer" containerID="5c90611e97642d06afcbae7dc9e52e2df50a2e1faaa225ef508016d671dd756f" Feb 27 16:54:55 crc kubenswrapper[4814]: I0227 16:54:55.594601 4814 scope.go:117] "RemoveContainer" containerID="cde917c2cc7628aa6eff69c0664fc22d8968642459b789c8e4e37989415ad0ab" Feb 27 16:54:55 crc kubenswrapper[4814]: I0227 16:54:55.648992 4814 scope.go:117] "RemoveContainer" containerID="0615c5bc9eee0c9e2440a32cd8f263d45418c571909c1a6d783625924de3e53c" Feb 27 16:54:55 crc kubenswrapper[4814]: I0227 16:54:55.696568 4814 scope.go:117] "RemoveContainer" containerID="efc4ef9a6d13d0d48a72b249a5e0c945110c36e452b9bc08e44c416a32b89469" Feb 27 16:54:55 crc kubenswrapper[4814]: I0227 16:54:55.722355 4814 scope.go:117] "RemoveContainer" containerID="64f19510352c40ef49f724f0e5f4340fbdf43eeba8b99ef52371658ff7008f4e" Feb 27 16:54:55 crc kubenswrapper[4814]: I0227 16:54:55.760273 4814 scope.go:117] "RemoveContainer" containerID="7b2a20b57b5372f102f10f54662a2ca73e2f1719ebc2a26b8a8d994150912faf" Feb 27 16:54:55 crc kubenswrapper[4814]: I0227 16:54:55.792786 4814 scope.go:117] "RemoveContainer" containerID="a0b6a424b296bbaed0cc99e6d686da04355e740cc540cb28cab975b8f3731f9a" Feb 27 16:54:55 crc kubenswrapper[4814]: I0227 16:54:55.826125 4814 scope.go:117] "RemoveContainer" containerID="0360f6ed9dd7bff9f9a14cba051f2c162805cc9ae3c673a435e3eb40c9f0f467" Feb 27 16:54:55 crc kubenswrapper[4814]: I0227 16:54:55.869450 4814 scope.go:117] "RemoveContainer" containerID="0b8580ebef7388d7c176924b75744c31dd6e900b6fe75afbfb66dc9414ef8eeb" Feb 27 16:54:55 crc kubenswrapper[4814]: I0227 16:54:55.908689 4814 scope.go:117] "RemoveContainer" containerID="2b112b351ca9b51bbab80bb80a2b1bd6887725e81a1392ec79e65bdd4d6dbc0b" Feb 27 16:54:59 crc kubenswrapper[4814]: I0227 16:54:59.488367 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:54:59 crc kubenswrapper[4814]: E0227 16:54:59.488969 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:55:02 crc kubenswrapper[4814]: I0227 16:55:02.851875 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:55:02 crc kubenswrapper[4814]: I0227 16:55:02.924384 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vzpq8"] Feb 27 16:55:03 crc kubenswrapper[4814]: I0227 16:55:03.250076 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vzpq8" podUID="7a144b53-17a0-4edd-a9fb-bb83cab694a5" containerName="registry-server" containerID="cri-o://48c3add4be409a817573417a17f4373ae8894a2ca1ad763cca808b40ec9a4e75" gracePeriod=2 Feb 27 16:55:03 crc kubenswrapper[4814]: I0227 16:55:03.734963 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:55:03 crc kubenswrapper[4814]: I0227 16:55:03.862519 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a144b53-17a0-4edd-a9fb-bb83cab694a5-utilities\") pod \"7a144b53-17a0-4edd-a9fb-bb83cab694a5\" (UID: \"7a144b53-17a0-4edd-a9fb-bb83cab694a5\") " Feb 27 16:55:03 crc kubenswrapper[4814]: I0227 16:55:03.863555 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb58l\" (UniqueName: \"kubernetes.io/projected/7a144b53-17a0-4edd-a9fb-bb83cab694a5-kube-api-access-wb58l\") pod \"7a144b53-17a0-4edd-a9fb-bb83cab694a5\" (UID: \"7a144b53-17a0-4edd-a9fb-bb83cab694a5\") " Feb 27 16:55:03 crc kubenswrapper[4814]: I0227 16:55:03.863778 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a144b53-17a0-4edd-a9fb-bb83cab694a5-catalog-content\") pod \"7a144b53-17a0-4edd-a9fb-bb83cab694a5\" (UID: \"7a144b53-17a0-4edd-a9fb-bb83cab694a5\") " Feb 27 16:55:03 crc kubenswrapper[4814]: I0227 16:55:03.864196 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a144b53-17a0-4edd-a9fb-bb83cab694a5-utilities" (OuterVolumeSpecName: "utilities") pod "7a144b53-17a0-4edd-a9fb-bb83cab694a5" (UID: "7a144b53-17a0-4edd-a9fb-bb83cab694a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:55:03 crc kubenswrapper[4814]: I0227 16:55:03.864704 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a144b53-17a0-4edd-a9fb-bb83cab694a5-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:55:03 crc kubenswrapper[4814]: I0227 16:55:03.871039 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a144b53-17a0-4edd-a9fb-bb83cab694a5-kube-api-access-wb58l" (OuterVolumeSpecName: "kube-api-access-wb58l") pod "7a144b53-17a0-4edd-a9fb-bb83cab694a5" (UID: "7a144b53-17a0-4edd-a9fb-bb83cab694a5"). InnerVolumeSpecName "kube-api-access-wb58l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:55:03 crc kubenswrapper[4814]: I0227 16:55:03.948384 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a144b53-17a0-4edd-a9fb-bb83cab694a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a144b53-17a0-4edd-a9fb-bb83cab694a5" (UID: "7a144b53-17a0-4edd-a9fb-bb83cab694a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:55:03 crc kubenswrapper[4814]: I0227 16:55:03.966982 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb58l\" (UniqueName: \"kubernetes.io/projected/7a144b53-17a0-4edd-a9fb-bb83cab694a5-kube-api-access-wb58l\") on node \"crc\" DevicePath \"\"" Feb 27 16:55:03 crc kubenswrapper[4814]: I0227 16:55:03.967011 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a144b53-17a0-4edd-a9fb-bb83cab694a5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:55:04 crc kubenswrapper[4814]: I0227 16:55:04.266544 4814 generic.go:334] "Generic (PLEG): container finished" podID="7a144b53-17a0-4edd-a9fb-bb83cab694a5" containerID="48c3add4be409a817573417a17f4373ae8894a2ca1ad763cca808b40ec9a4e75" exitCode=0 Feb 27 16:55:04 crc kubenswrapper[4814]: I0227 16:55:04.266607 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzpq8" event={"ID":"7a144b53-17a0-4edd-a9fb-bb83cab694a5","Type":"ContainerDied","Data":"48c3add4be409a817573417a17f4373ae8894a2ca1ad763cca808b40ec9a4e75"} Feb 27 16:55:04 crc kubenswrapper[4814]: I0227 16:55:04.266659 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzpq8" event={"ID":"7a144b53-17a0-4edd-a9fb-bb83cab694a5","Type":"ContainerDied","Data":"7a9b591e6416a05dcc174e054e043ea0be660442624de4589099c5f24454c94a"} Feb 27 16:55:04 crc kubenswrapper[4814]: I0227 16:55:04.266663 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vzpq8" Feb 27 16:55:04 crc kubenswrapper[4814]: I0227 16:55:04.266690 4814 scope.go:117] "RemoveContainer" containerID="48c3add4be409a817573417a17f4373ae8894a2ca1ad763cca808b40ec9a4e75" Feb 27 16:55:04 crc kubenswrapper[4814]: I0227 16:55:04.298240 4814 scope.go:117] "RemoveContainer" containerID="f9094016e0692db1d09824f955d031d5c4fbd67e11c8e567b466fcb444f6f8af" Feb 27 16:55:04 crc kubenswrapper[4814]: I0227 16:55:04.330732 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vzpq8"] Feb 27 16:55:04 crc kubenswrapper[4814]: I0227 16:55:04.339930 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vzpq8"] Feb 27 16:55:04 crc kubenswrapper[4814]: I0227 16:55:04.348667 4814 scope.go:117] "RemoveContainer" containerID="948b40794b5f8529e7c6dbd13cf3768c6edbd4655812250602b46ef447b813d9" Feb 27 16:55:04 crc kubenswrapper[4814]: I0227 16:55:04.391543 4814 scope.go:117] "RemoveContainer" containerID="48c3add4be409a817573417a17f4373ae8894a2ca1ad763cca808b40ec9a4e75" Feb 27 16:55:04 crc kubenswrapper[4814]: E0227 16:55:04.393249 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48c3add4be409a817573417a17f4373ae8894a2ca1ad763cca808b40ec9a4e75\": container with ID starting with 48c3add4be409a817573417a17f4373ae8894a2ca1ad763cca808b40ec9a4e75 not found: ID does not exist" containerID="48c3add4be409a817573417a17f4373ae8894a2ca1ad763cca808b40ec9a4e75" Feb 27 16:55:04 crc kubenswrapper[4814]: I0227 16:55:04.393344 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48c3add4be409a817573417a17f4373ae8894a2ca1ad763cca808b40ec9a4e75"} err="failed to get container status \"48c3add4be409a817573417a17f4373ae8894a2ca1ad763cca808b40ec9a4e75\": rpc error: code = NotFound desc = could not find container \"48c3add4be409a817573417a17f4373ae8894a2ca1ad763cca808b40ec9a4e75\": container with ID starting with 48c3add4be409a817573417a17f4373ae8894a2ca1ad763cca808b40ec9a4e75 not found: ID does not exist" Feb 27 16:55:04 crc kubenswrapper[4814]: I0227 16:55:04.393384 4814 scope.go:117] "RemoveContainer" containerID="f9094016e0692db1d09824f955d031d5c4fbd67e11c8e567b466fcb444f6f8af" Feb 27 16:55:04 crc kubenswrapper[4814]: E0227 16:55:04.393751 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9094016e0692db1d09824f955d031d5c4fbd67e11c8e567b466fcb444f6f8af\": container with ID starting with f9094016e0692db1d09824f955d031d5c4fbd67e11c8e567b466fcb444f6f8af not found: ID does not exist" containerID="f9094016e0692db1d09824f955d031d5c4fbd67e11c8e567b466fcb444f6f8af" Feb 27 16:55:04 crc kubenswrapper[4814]: I0227 16:55:04.393796 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9094016e0692db1d09824f955d031d5c4fbd67e11c8e567b466fcb444f6f8af"} err="failed to get container status \"f9094016e0692db1d09824f955d031d5c4fbd67e11c8e567b466fcb444f6f8af\": rpc error: code = NotFound desc = could not find container \"f9094016e0692db1d09824f955d031d5c4fbd67e11c8e567b466fcb444f6f8af\": container with ID starting with f9094016e0692db1d09824f955d031d5c4fbd67e11c8e567b466fcb444f6f8af not found: ID does not exist" Feb 27 16:55:04 crc kubenswrapper[4814]: I0227 16:55:04.393829 4814 scope.go:117] "RemoveContainer" containerID="948b40794b5f8529e7c6dbd13cf3768c6edbd4655812250602b46ef447b813d9" Feb 27 16:55:04 crc kubenswrapper[4814]: E0227 16:55:04.394099 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"948b40794b5f8529e7c6dbd13cf3768c6edbd4655812250602b46ef447b813d9\": container with ID starting with 948b40794b5f8529e7c6dbd13cf3768c6edbd4655812250602b46ef447b813d9 not found: ID does not exist" containerID="948b40794b5f8529e7c6dbd13cf3768c6edbd4655812250602b46ef447b813d9" Feb 27 16:55:04 crc kubenswrapper[4814]: I0227 16:55:04.394138 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"948b40794b5f8529e7c6dbd13cf3768c6edbd4655812250602b46ef447b813d9"} err="failed to get container status \"948b40794b5f8529e7c6dbd13cf3768c6edbd4655812250602b46ef447b813d9\": rpc error: code = NotFound desc = could not find container \"948b40794b5f8529e7c6dbd13cf3768c6edbd4655812250602b46ef447b813d9\": container with ID starting with 948b40794b5f8529e7c6dbd13cf3768c6edbd4655812250602b46ef447b813d9 not found: ID does not exist" Feb 27 16:55:04 crc kubenswrapper[4814]: I0227 16:55:04.500147 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a144b53-17a0-4edd-a9fb-bb83cab694a5" path="/var/lib/kubelet/pods/7a144b53-17a0-4edd-a9fb-bb83cab694a5/volumes" Feb 27 16:55:11 crc kubenswrapper[4814]: I0227 16:55:11.488195 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:55:11 crc kubenswrapper[4814]: E0227 16:55:11.489420 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:55:20 crc kubenswrapper[4814]: I0227 16:55:20.472608 4814 generic.go:334] "Generic (PLEG): container finished" podID="f4d8ff3a-2b21-45ab-8332-5bedf355c09c" containerID="4e09917a07d52e2ec81a2c1184e5a97f383bea284ad0b6cc93b886849477685c" exitCode=0 Feb 27 16:55:20 crc kubenswrapper[4814]: I0227 16:55:20.472701 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" event={"ID":"f4d8ff3a-2b21-45ab-8332-5bedf355c09c","Type":"ContainerDied","Data":"4e09917a07d52e2ec81a2c1184e5a97f383bea284ad0b6cc93b886849477685c"} Feb 27 16:55:21 crc kubenswrapper[4814]: I0227 16:55:21.974797 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" Feb 27 16:55:21 crc kubenswrapper[4814]: I0227 16:55:21.987425 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-ssh-key-openstack-edpm-ipam\") pod \"f4d8ff3a-2b21-45ab-8332-5bedf355c09c\" (UID: \"f4d8ff3a-2b21-45ab-8332-5bedf355c09c\") " Feb 27 16:55:21 crc kubenswrapper[4814]: I0227 16:55:21.987475 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-inventory\") pod \"f4d8ff3a-2b21-45ab-8332-5bedf355c09c\" (UID: \"f4d8ff3a-2b21-45ab-8332-5bedf355c09c\") " Feb 27 16:55:21 crc kubenswrapper[4814]: I0227 16:55:21.987507 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmjln\" (UniqueName: \"kubernetes.io/projected/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-kube-api-access-nmjln\") pod \"f4d8ff3a-2b21-45ab-8332-5bedf355c09c\" (UID: \"f4d8ff3a-2b21-45ab-8332-5bedf355c09c\") " Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.000449 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-kube-api-access-nmjln" (OuterVolumeSpecName: "kube-api-access-nmjln") pod "f4d8ff3a-2b21-45ab-8332-5bedf355c09c" (UID: "f4d8ff3a-2b21-45ab-8332-5bedf355c09c"). InnerVolumeSpecName "kube-api-access-nmjln". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.044474 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-inventory" (OuterVolumeSpecName: "inventory") pod "f4d8ff3a-2b21-45ab-8332-5bedf355c09c" (UID: "f4d8ff3a-2b21-45ab-8332-5bedf355c09c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.048554 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f4d8ff3a-2b21-45ab-8332-5bedf355c09c" (UID: "f4d8ff3a-2b21-45ab-8332-5bedf355c09c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.071414 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jjmg8"] Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.076901 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jjmg8"] Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.096200 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.096236 4814 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.098399 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmjln\" (UniqueName: \"kubernetes.io/projected/f4d8ff3a-2b21-45ab-8332-5bedf355c09c-kube-api-access-nmjln\") on node \"crc\" DevicePath \"\"" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.498536 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.540126 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="827ecd26-d313-4fb7-86bb-03e6e777f5a7" path="/var/lib/kubelet/pods/827ecd26-d313-4fb7-86bb-03e6e777f5a7/volumes" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.541876 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sqtqt" event={"ID":"f4d8ff3a-2b21-45ab-8332-5bedf355c09c","Type":"ContainerDied","Data":"4262886ad9da8c01017c55ff662a675d638dc8e1935a9d8ecadcb39df7ec1446"} Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.541931 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4262886ad9da8c01017c55ff662a675d638dc8e1935a9d8ecadcb39df7ec1446" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.640618 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x"] Feb 27 16:55:22 crc kubenswrapper[4814]: E0227 16:55:22.641025 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4d8ff3a-2b21-45ab-8332-5bedf355c09c" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.641045 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4d8ff3a-2b21-45ab-8332-5bedf355c09c" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 27 16:55:22 crc kubenswrapper[4814]: E0227 16:55:22.641068 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a144b53-17a0-4edd-a9fb-bb83cab694a5" containerName="registry-server" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.641076 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a144b53-17a0-4edd-a9fb-bb83cab694a5" containerName="registry-server" Feb 27 16:55:22 crc kubenswrapper[4814]: E0227 16:55:22.641092 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a144b53-17a0-4edd-a9fb-bb83cab694a5" containerName="extract-utilities" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.641100 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a144b53-17a0-4edd-a9fb-bb83cab694a5" containerName="extract-utilities" Feb 27 16:55:22 crc kubenswrapper[4814]: E0227 16:55:22.641118 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a144b53-17a0-4edd-a9fb-bb83cab694a5" containerName="extract-content" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.641125 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a144b53-17a0-4edd-a9fb-bb83cab694a5" containerName="extract-content" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.641336 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4d8ff3a-2b21-45ab-8332-5bedf355c09c" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.641353 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a144b53-17a0-4edd-a9fb-bb83cab694a5" containerName="registry-server" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.642032 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.644964 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.645292 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.645584 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.645747 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.668122 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x"] Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.810767 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3d27f572-8f1b-4507-a482-f3ce12031e76-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-59k2x\" (UID: \"3d27f572-8f1b-4507-a482-f3ce12031e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.810932 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4dl6\" (UniqueName: \"kubernetes.io/projected/3d27f572-8f1b-4507-a482-f3ce12031e76-kube-api-access-m4dl6\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-59k2x\" (UID: \"3d27f572-8f1b-4507-a482-f3ce12031e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.810962 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d27f572-8f1b-4507-a482-f3ce12031e76-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-59k2x\" (UID: \"3d27f572-8f1b-4507-a482-f3ce12031e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.912918 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d27f572-8f1b-4507-a482-f3ce12031e76-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-59k2x\" (UID: \"3d27f572-8f1b-4507-a482-f3ce12031e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.912985 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4dl6\" (UniqueName: \"kubernetes.io/projected/3d27f572-8f1b-4507-a482-f3ce12031e76-kube-api-access-m4dl6\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-59k2x\" (UID: \"3d27f572-8f1b-4507-a482-f3ce12031e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.913131 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3d27f572-8f1b-4507-a482-f3ce12031e76-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-59k2x\" (UID: \"3d27f572-8f1b-4507-a482-f3ce12031e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.925307 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3d27f572-8f1b-4507-a482-f3ce12031e76-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-59k2x\" (UID: \"3d27f572-8f1b-4507-a482-f3ce12031e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.929181 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d27f572-8f1b-4507-a482-f3ce12031e76-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-59k2x\" (UID: \"3d27f572-8f1b-4507-a482-f3ce12031e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.936528 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4dl6\" (UniqueName: \"kubernetes.io/projected/3d27f572-8f1b-4507-a482-f3ce12031e76-kube-api-access-m4dl6\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-59k2x\" (UID: \"3d27f572-8f1b-4507-a482-f3ce12031e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" Feb 27 16:55:22 crc kubenswrapper[4814]: I0227 16:55:22.971533 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" Feb 27 16:55:23 crc kubenswrapper[4814]: I0227 16:55:23.407854 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x"] Feb 27 16:55:23 crc kubenswrapper[4814]: I0227 16:55:23.508026 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" event={"ID":"3d27f572-8f1b-4507-a482-f3ce12031e76","Type":"ContainerStarted","Data":"53c0cf3e95fe6442d2a68d7c7046cf35b074b15b57b39783939cd23e5df182ca"} Feb 27 16:55:24 crc kubenswrapper[4814]: I0227 16:55:24.547972 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" event={"ID":"3d27f572-8f1b-4507-a482-f3ce12031e76","Type":"ContainerStarted","Data":"caf6c1d21711ba0fa06a8f77f3d866689eb0a2afc795511bfba7feba94001092"} Feb 27 16:55:24 crc kubenswrapper[4814]: I0227 16:55:24.574042 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" podStartSLOduration=2.095494007 podStartE2EDuration="2.574022304s" podCreationTimestamp="2026-02-27 16:55:22 +0000 UTC" firstStartedPulling="2026-02-27 16:55:23.411358716 +0000 UTC m=+1935.863983546" lastFinishedPulling="2026-02-27 16:55:23.889886983 +0000 UTC m=+1936.342511843" observedRunningTime="2026-02-27 16:55:24.564856944 +0000 UTC m=+1937.017481764" watchObservedRunningTime="2026-02-27 16:55:24.574022304 +0000 UTC m=+1937.026647134" Feb 27 16:55:26 crc kubenswrapper[4814]: I0227 16:55:26.488371 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:55:26 crc kubenswrapper[4814]: E0227 16:55:26.489149 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:55:41 crc kubenswrapper[4814]: I0227 16:55:41.488194 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:55:41 crc kubenswrapper[4814]: E0227 16:55:41.489375 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:55:45 crc kubenswrapper[4814]: I0227 16:55:45.051456 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-52n2r"] Feb 27 16:55:45 crc kubenswrapper[4814]: I0227 16:55:45.065993 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-52n2r"] Feb 27 16:55:46 crc kubenswrapper[4814]: I0227 16:55:46.505932 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66b6902c-c560-43ba-8bf6-b162da3931df" path="/var/lib/kubelet/pods/66b6902c-c560-43ba-8bf6-b162da3931df/volumes" Feb 27 16:55:47 crc kubenswrapper[4814]: I0227 16:55:47.031246 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qm8cq"] Feb 27 16:55:47 crc kubenswrapper[4814]: I0227 16:55:47.040083 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qm8cq"] Feb 27 16:55:48 crc kubenswrapper[4814]: I0227 16:55:48.504123 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8be859d-3be2-4c3f-b10e-b0d04d8fe14d" path="/var/lib/kubelet/pods/f8be859d-3be2-4c3f-b10e-b0d04d8fe14d/volumes" Feb 27 16:55:56 crc kubenswrapper[4814]: I0227 16:55:56.222603 4814 scope.go:117] "RemoveContainer" containerID="ca4d6256ededf270e247f60cc5ccd983bac59c8fc25d2020e5e5bbdc188cf590" Feb 27 16:55:56 crc kubenswrapper[4814]: I0227 16:55:56.267537 4814 scope.go:117] "RemoveContainer" containerID="5f2cc4f12adcccdec8db26f18084d239bc8397263cda448b3eb475afe5553635" Feb 27 16:55:56 crc kubenswrapper[4814]: I0227 16:55:56.312392 4814 scope.go:117] "RemoveContainer" containerID="838f5dea6dde2748cbb8c5f618de14790336077931da329733c57a834e12c718" Feb 27 16:55:56 crc kubenswrapper[4814]: I0227 16:55:56.488044 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:55:56 crc kubenswrapper[4814]: E0227 16:55:56.488415 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:56:00 crc kubenswrapper[4814]: I0227 16:56:00.161831 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536856-9mgrh"] Feb 27 16:56:00 crc kubenswrapper[4814]: I0227 16:56:00.165458 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536856-9mgrh" Feb 27 16:56:00 crc kubenswrapper[4814]: I0227 16:56:00.168811 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:56:00 crc kubenswrapper[4814]: I0227 16:56:00.169075 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:56:00 crc kubenswrapper[4814]: I0227 16:56:00.169111 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 16:56:00 crc kubenswrapper[4814]: I0227 16:56:00.209977 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536856-9mgrh"] Feb 27 16:56:00 crc kubenswrapper[4814]: I0227 16:56:00.269543 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76962\" (UniqueName: \"kubernetes.io/projected/3c9122d2-7427-4dd8-974d-38cfce5c3cad-kube-api-access-76962\") pod \"auto-csr-approver-29536856-9mgrh\" (UID: \"3c9122d2-7427-4dd8-974d-38cfce5c3cad\") " pod="openshift-infra/auto-csr-approver-29536856-9mgrh" Feb 27 16:56:00 crc kubenswrapper[4814]: I0227 16:56:00.371511 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76962\" (UniqueName: \"kubernetes.io/projected/3c9122d2-7427-4dd8-974d-38cfce5c3cad-kube-api-access-76962\") pod \"auto-csr-approver-29536856-9mgrh\" (UID: \"3c9122d2-7427-4dd8-974d-38cfce5c3cad\") " pod="openshift-infra/auto-csr-approver-29536856-9mgrh" Feb 27 16:56:00 crc kubenswrapper[4814]: I0227 16:56:00.408961 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76962\" (UniqueName: \"kubernetes.io/projected/3c9122d2-7427-4dd8-974d-38cfce5c3cad-kube-api-access-76962\") pod \"auto-csr-approver-29536856-9mgrh\" (UID: \"3c9122d2-7427-4dd8-974d-38cfce5c3cad\") " pod="openshift-infra/auto-csr-approver-29536856-9mgrh" Feb 27 16:56:00 crc kubenswrapper[4814]: I0227 16:56:00.516655 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536856-9mgrh" Feb 27 16:56:01 crc kubenswrapper[4814]: I0227 16:56:01.002908 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536856-9mgrh"] Feb 27 16:56:01 crc kubenswrapper[4814]: W0227 16:56:01.009734 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c9122d2_7427_4dd8_974d_38cfce5c3cad.slice/crio-0249d620afca8cd9f6e4904041ba3e72cb4928f3f723afaf8b35cd50da4dd294 WatchSource:0}: Error finding container 0249d620afca8cd9f6e4904041ba3e72cb4928f3f723afaf8b35cd50da4dd294: Status 404 returned error can't find the container with id 0249d620afca8cd9f6e4904041ba3e72cb4928f3f723afaf8b35cd50da4dd294 Feb 27 16:56:01 crc kubenswrapper[4814]: I0227 16:56:01.953836 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536856-9mgrh" event={"ID":"3c9122d2-7427-4dd8-974d-38cfce5c3cad","Type":"ContainerStarted","Data":"0249d620afca8cd9f6e4904041ba3e72cb4928f3f723afaf8b35cd50da4dd294"} Feb 27 16:56:02 crc kubenswrapper[4814]: I0227 16:56:02.967003 4814 generic.go:334] "Generic (PLEG): container finished" podID="3c9122d2-7427-4dd8-974d-38cfce5c3cad" containerID="3e6fd5aed07cd58e1536aaf9534be94f5bebd7fcf616b732cb96128307fe4245" exitCode=0 Feb 27 16:56:02 crc kubenswrapper[4814]: I0227 16:56:02.967085 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536856-9mgrh" event={"ID":"3c9122d2-7427-4dd8-974d-38cfce5c3cad","Type":"ContainerDied","Data":"3e6fd5aed07cd58e1536aaf9534be94f5bebd7fcf616b732cb96128307fe4245"} Feb 27 16:56:04 crc kubenswrapper[4814]: I0227 16:56:04.336540 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536856-9mgrh" Feb 27 16:56:04 crc kubenswrapper[4814]: I0227 16:56:04.463160 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76962\" (UniqueName: \"kubernetes.io/projected/3c9122d2-7427-4dd8-974d-38cfce5c3cad-kube-api-access-76962\") pod \"3c9122d2-7427-4dd8-974d-38cfce5c3cad\" (UID: \"3c9122d2-7427-4dd8-974d-38cfce5c3cad\") " Feb 27 16:56:04 crc kubenswrapper[4814]: I0227 16:56:04.476592 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c9122d2-7427-4dd8-974d-38cfce5c3cad-kube-api-access-76962" (OuterVolumeSpecName: "kube-api-access-76962") pod "3c9122d2-7427-4dd8-974d-38cfce5c3cad" (UID: "3c9122d2-7427-4dd8-974d-38cfce5c3cad"). InnerVolumeSpecName "kube-api-access-76962". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:56:04 crc kubenswrapper[4814]: I0227 16:56:04.566511 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76962\" (UniqueName: \"kubernetes.io/projected/3c9122d2-7427-4dd8-974d-38cfce5c3cad-kube-api-access-76962\") on node \"crc\" DevicePath \"\"" Feb 27 16:56:04 crc kubenswrapper[4814]: I0227 16:56:04.991929 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536856-9mgrh" event={"ID":"3c9122d2-7427-4dd8-974d-38cfce5c3cad","Type":"ContainerDied","Data":"0249d620afca8cd9f6e4904041ba3e72cb4928f3f723afaf8b35cd50da4dd294"} Feb 27 16:56:04 crc kubenswrapper[4814]: I0227 16:56:04.991991 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0249d620afca8cd9f6e4904041ba3e72cb4928f3f723afaf8b35cd50da4dd294" Feb 27 16:56:04 crc kubenswrapper[4814]: I0227 16:56:04.992012 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536856-9mgrh" Feb 27 16:56:05 crc kubenswrapper[4814]: I0227 16:56:05.438365 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536850-xqntl"] Feb 27 16:56:05 crc kubenswrapper[4814]: I0227 16:56:05.453158 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536850-xqntl"] Feb 27 16:56:06 crc kubenswrapper[4814]: I0227 16:56:06.499927 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b661ed9f-ec8e-42bd-86eb-3c4da181ced4" path="/var/lib/kubelet/pods/b661ed9f-ec8e-42bd-86eb-3c4da181ced4/volumes" Feb 27 16:56:08 crc kubenswrapper[4814]: I0227 16:56:08.501642 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:56:08 crc kubenswrapper[4814]: E0227 16:56:08.502140 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:56:15 crc kubenswrapper[4814]: I0227 16:56:15.098405 4814 generic.go:334] "Generic (PLEG): container finished" podID="3d27f572-8f1b-4507-a482-f3ce12031e76" containerID="caf6c1d21711ba0fa06a8f77f3d866689eb0a2afc795511bfba7feba94001092" exitCode=0 Feb 27 16:56:15 crc kubenswrapper[4814]: I0227 16:56:15.098498 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" event={"ID":"3d27f572-8f1b-4507-a482-f3ce12031e76","Type":"ContainerDied","Data":"caf6c1d21711ba0fa06a8f77f3d866689eb0a2afc795511bfba7feba94001092"} Feb 27 16:56:16 crc kubenswrapper[4814]: I0227 16:56:16.832413 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" Feb 27 16:56:16 crc kubenswrapper[4814]: I0227 16:56:16.942285 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4dl6\" (UniqueName: \"kubernetes.io/projected/3d27f572-8f1b-4507-a482-f3ce12031e76-kube-api-access-m4dl6\") pod \"3d27f572-8f1b-4507-a482-f3ce12031e76\" (UID: \"3d27f572-8f1b-4507-a482-f3ce12031e76\") " Feb 27 16:56:16 crc kubenswrapper[4814]: I0227 16:56:16.942660 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d27f572-8f1b-4507-a482-f3ce12031e76-inventory\") pod \"3d27f572-8f1b-4507-a482-f3ce12031e76\" (UID: \"3d27f572-8f1b-4507-a482-f3ce12031e76\") " Feb 27 16:56:16 crc kubenswrapper[4814]: I0227 16:56:16.942877 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3d27f572-8f1b-4507-a482-f3ce12031e76-ssh-key-openstack-edpm-ipam\") pod \"3d27f572-8f1b-4507-a482-f3ce12031e76\" (UID: \"3d27f572-8f1b-4507-a482-f3ce12031e76\") " Feb 27 16:56:16 crc kubenswrapper[4814]: I0227 16:56:16.958487 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d27f572-8f1b-4507-a482-f3ce12031e76-kube-api-access-m4dl6" (OuterVolumeSpecName: "kube-api-access-m4dl6") pod "3d27f572-8f1b-4507-a482-f3ce12031e76" (UID: "3d27f572-8f1b-4507-a482-f3ce12031e76"). InnerVolumeSpecName "kube-api-access-m4dl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:56:16 crc kubenswrapper[4814]: I0227 16:56:16.974296 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d27f572-8f1b-4507-a482-f3ce12031e76-inventory" (OuterVolumeSpecName: "inventory") pod "3d27f572-8f1b-4507-a482-f3ce12031e76" (UID: "3d27f572-8f1b-4507-a482-f3ce12031e76"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:56:16 crc kubenswrapper[4814]: I0227 16:56:16.985971 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d27f572-8f1b-4507-a482-f3ce12031e76-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3d27f572-8f1b-4507-a482-f3ce12031e76" (UID: "3d27f572-8f1b-4507-a482-f3ce12031e76"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.045613 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4dl6\" (UniqueName: \"kubernetes.io/projected/3d27f572-8f1b-4507-a482-f3ce12031e76-kube-api-access-m4dl6\") on node \"crc\" DevicePath \"\"" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.045653 4814 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d27f572-8f1b-4507-a482-f3ce12031e76-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.045666 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3d27f572-8f1b-4507-a482-f3ce12031e76-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.136952 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" event={"ID":"3d27f572-8f1b-4507-a482-f3ce12031e76","Type":"ContainerDied","Data":"53c0cf3e95fe6442d2a68d7c7046cf35b074b15b57b39783939cd23e5df182ca"} Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.137001 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53c0cf3e95fe6442d2a68d7c7046cf35b074b15b57b39783939cd23e5df182ca" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.137044 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-59k2x" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.243631 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-kz77t"] Feb 27 16:56:17 crc kubenswrapper[4814]: E0227 16:56:17.244206 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d27f572-8f1b-4507-a482-f3ce12031e76" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.244235 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d27f572-8f1b-4507-a482-f3ce12031e76" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 27 16:56:17 crc kubenswrapper[4814]: E0227 16:56:17.244321 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c9122d2-7427-4dd8-974d-38cfce5c3cad" containerName="oc" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.244335 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c9122d2-7427-4dd8-974d-38cfce5c3cad" containerName="oc" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.244682 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d27f572-8f1b-4507-a482-f3ce12031e76" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.244717 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c9122d2-7427-4dd8-974d-38cfce5c3cad" containerName="oc" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.245694 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.249331 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.249978 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.250227 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.250759 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.258919 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-kz77t"] Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.354152 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwrc8\" (UniqueName: \"kubernetes.io/projected/6fddd1b5-7d4c-4986-b192-f06f528c4b64-kube-api-access-lwrc8\") pod \"ssh-known-hosts-edpm-deployment-kz77t\" (UID: \"6fddd1b5-7d4c-4986-b192-f06f528c4b64\") " pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.354595 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6fddd1b5-7d4c-4986-b192-f06f528c4b64-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-kz77t\" (UID: \"6fddd1b5-7d4c-4986-b192-f06f528c4b64\") " pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.354834 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6fddd1b5-7d4c-4986-b192-f06f528c4b64-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-kz77t\" (UID: \"6fddd1b5-7d4c-4986-b192-f06f528c4b64\") " pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.457374 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6fddd1b5-7d4c-4986-b192-f06f528c4b64-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-kz77t\" (UID: \"6fddd1b5-7d4c-4986-b192-f06f528c4b64\") " pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.457476 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwrc8\" (UniqueName: \"kubernetes.io/projected/6fddd1b5-7d4c-4986-b192-f06f528c4b64-kube-api-access-lwrc8\") pod \"ssh-known-hosts-edpm-deployment-kz77t\" (UID: \"6fddd1b5-7d4c-4986-b192-f06f528c4b64\") " pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.458013 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6fddd1b5-7d4c-4986-b192-f06f528c4b64-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-kz77t\" (UID: \"6fddd1b5-7d4c-4986-b192-f06f528c4b64\") " pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.467974 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6fddd1b5-7d4c-4986-b192-f06f528c4b64-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-kz77t\" (UID: \"6fddd1b5-7d4c-4986-b192-f06f528c4b64\") " pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.471692 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6fddd1b5-7d4c-4986-b192-f06f528c4b64-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-kz77t\" (UID: \"6fddd1b5-7d4c-4986-b192-f06f528c4b64\") " pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.490152 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwrc8\" (UniqueName: \"kubernetes.io/projected/6fddd1b5-7d4c-4986-b192-f06f528c4b64-kube-api-access-lwrc8\") pod \"ssh-known-hosts-edpm-deployment-kz77t\" (UID: \"6fddd1b5-7d4c-4986-b192-f06f528c4b64\") " pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.579519 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" Feb 27 16:56:17 crc kubenswrapper[4814]: I0227 16:56:17.956405 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-kz77t"] Feb 27 16:56:18 crc kubenswrapper[4814]: I0227 16:56:18.151871 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" event={"ID":"6fddd1b5-7d4c-4986-b192-f06f528c4b64","Type":"ContainerStarted","Data":"2232f034ff276e0352d91ee89e60dad0c39a5b285a60bbcba57860690a6e0288"} Feb 27 16:56:19 crc kubenswrapper[4814]: I0227 16:56:19.169192 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" event={"ID":"6fddd1b5-7d4c-4986-b192-f06f528c4b64","Type":"ContainerStarted","Data":"138a39b0831d80a21e2cbd37c935b3b9f431a3870eb2aca47d488b1332793702"} Feb 27 16:56:19 crc kubenswrapper[4814]: I0227 16:56:19.198412 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" podStartSLOduration=1.763033063 podStartE2EDuration="2.19837601s" podCreationTimestamp="2026-02-27 16:56:17 +0000 UTC" firstStartedPulling="2026-02-27 16:56:17.965140055 +0000 UTC m=+1990.417764895" lastFinishedPulling="2026-02-27 16:56:18.400482962 +0000 UTC m=+1990.853107842" observedRunningTime="2026-02-27 16:56:19.195996568 +0000 UTC m=+1991.648621428" watchObservedRunningTime="2026-02-27 16:56:19.19837601 +0000 UTC m=+1991.651000870" Feb 27 16:56:19 crc kubenswrapper[4814]: I0227 16:56:19.488595 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:56:19 crc kubenswrapper[4814]: E0227 16:56:19.489105 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:56:26 crc kubenswrapper[4814]: I0227 16:56:26.244396 4814 generic.go:334] "Generic (PLEG): container finished" podID="6fddd1b5-7d4c-4986-b192-f06f528c4b64" containerID="138a39b0831d80a21e2cbd37c935b3b9f431a3870eb2aca47d488b1332793702" exitCode=0 Feb 27 16:56:26 crc kubenswrapper[4814]: I0227 16:56:26.244515 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" event={"ID":"6fddd1b5-7d4c-4986-b192-f06f528c4b64","Type":"ContainerDied","Data":"138a39b0831d80a21e2cbd37c935b3b9f431a3870eb2aca47d488b1332793702"} Feb 27 16:56:27 crc kubenswrapper[4814]: I0227 16:56:27.803621 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" Feb 27 16:56:27 crc kubenswrapper[4814]: I0227 16:56:27.886412 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6fddd1b5-7d4c-4986-b192-f06f528c4b64-ssh-key-openstack-edpm-ipam\") pod \"6fddd1b5-7d4c-4986-b192-f06f528c4b64\" (UID: \"6fddd1b5-7d4c-4986-b192-f06f528c4b64\") " Feb 27 16:56:27 crc kubenswrapper[4814]: I0227 16:56:27.887299 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwrc8\" (UniqueName: \"kubernetes.io/projected/6fddd1b5-7d4c-4986-b192-f06f528c4b64-kube-api-access-lwrc8\") pod \"6fddd1b5-7d4c-4986-b192-f06f528c4b64\" (UID: \"6fddd1b5-7d4c-4986-b192-f06f528c4b64\") " Feb 27 16:56:27 crc kubenswrapper[4814]: I0227 16:56:27.887417 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6fddd1b5-7d4c-4986-b192-f06f528c4b64-inventory-0\") pod \"6fddd1b5-7d4c-4986-b192-f06f528c4b64\" (UID: \"6fddd1b5-7d4c-4986-b192-f06f528c4b64\") " Feb 27 16:56:27 crc kubenswrapper[4814]: I0227 16:56:27.900920 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fddd1b5-7d4c-4986-b192-f06f528c4b64-kube-api-access-lwrc8" (OuterVolumeSpecName: "kube-api-access-lwrc8") pod "6fddd1b5-7d4c-4986-b192-f06f528c4b64" (UID: "6fddd1b5-7d4c-4986-b192-f06f528c4b64"). InnerVolumeSpecName "kube-api-access-lwrc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:56:27 crc kubenswrapper[4814]: I0227 16:56:27.944123 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fddd1b5-7d4c-4986-b192-f06f528c4b64-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "6fddd1b5-7d4c-4986-b192-f06f528c4b64" (UID: "6fddd1b5-7d4c-4986-b192-f06f528c4b64"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:56:27 crc kubenswrapper[4814]: I0227 16:56:27.956962 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fddd1b5-7d4c-4986-b192-f06f528c4b64-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6fddd1b5-7d4c-4986-b192-f06f528c4b64" (UID: "6fddd1b5-7d4c-4986-b192-f06f528c4b64"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:56:27 crc kubenswrapper[4814]: I0227 16:56:27.989030 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6fddd1b5-7d4c-4986-b192-f06f528c4b64-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 16:56:27 crc kubenswrapper[4814]: I0227 16:56:27.989067 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwrc8\" (UniqueName: \"kubernetes.io/projected/6fddd1b5-7d4c-4986-b192-f06f528c4b64-kube-api-access-lwrc8\") on node \"crc\" DevicePath \"\"" Feb 27 16:56:27 crc kubenswrapper[4814]: I0227 16:56:27.989079 4814 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6fddd1b5-7d4c-4986-b192-f06f528c4b64-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.268320 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" event={"ID":"6fddd1b5-7d4c-4986-b192-f06f528c4b64","Type":"ContainerDied","Data":"2232f034ff276e0352d91ee89e60dad0c39a5b285a60bbcba57860690a6e0288"} Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.268388 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2232f034ff276e0352d91ee89e60dad0c39a5b285a60bbcba57860690a6e0288" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.268405 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-kz77t" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.351739 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk"] Feb 27 16:56:28 crc kubenswrapper[4814]: E0227 16:56:28.352160 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fddd1b5-7d4c-4986-b192-f06f528c4b64" containerName="ssh-known-hosts-edpm-deployment" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.352180 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fddd1b5-7d4c-4986-b192-f06f528c4b64" containerName="ssh-known-hosts-edpm-deployment" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.352452 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fddd1b5-7d4c-4986-b192-f06f528c4b64" containerName="ssh-known-hosts-edpm-deployment" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.353158 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.356480 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.357908 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.358476 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.358476 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.373748 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk"] Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.396158 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/223a940a-b149-4b9a-a328-411a730cf5fb-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s5rjk\" (UID: \"223a940a-b149-4b9a-a328-411a730cf5fb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.396338 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/223a940a-b149-4b9a-a328-411a730cf5fb-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s5rjk\" (UID: \"223a940a-b149-4b9a-a328-411a730cf5fb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.396424 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f6r4\" (UniqueName: \"kubernetes.io/projected/223a940a-b149-4b9a-a328-411a730cf5fb-kube-api-access-9f6r4\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s5rjk\" (UID: \"223a940a-b149-4b9a-a328-411a730cf5fb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.498067 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/223a940a-b149-4b9a-a328-411a730cf5fb-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s5rjk\" (UID: \"223a940a-b149-4b9a-a328-411a730cf5fb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.498180 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/223a940a-b149-4b9a-a328-411a730cf5fb-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s5rjk\" (UID: \"223a940a-b149-4b9a-a328-411a730cf5fb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.498333 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f6r4\" (UniqueName: \"kubernetes.io/projected/223a940a-b149-4b9a-a328-411a730cf5fb-kube-api-access-9f6r4\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s5rjk\" (UID: \"223a940a-b149-4b9a-a328-411a730cf5fb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.503333 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/223a940a-b149-4b9a-a328-411a730cf5fb-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s5rjk\" (UID: \"223a940a-b149-4b9a-a328-411a730cf5fb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.504146 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/223a940a-b149-4b9a-a328-411a730cf5fb-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s5rjk\" (UID: \"223a940a-b149-4b9a-a328-411a730cf5fb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.528211 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f6r4\" (UniqueName: \"kubernetes.io/projected/223a940a-b149-4b9a-a328-411a730cf5fb-kube-api-access-9f6r4\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s5rjk\" (UID: \"223a940a-b149-4b9a-a328-411a730cf5fb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" Feb 27 16:56:28 crc kubenswrapper[4814]: I0227 16:56:28.681606 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" Feb 27 16:56:29 crc kubenswrapper[4814]: I0227 16:56:29.053421 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-l695f"] Feb 27 16:56:29 crc kubenswrapper[4814]: I0227 16:56:29.067332 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-l695f"] Feb 27 16:56:29 crc kubenswrapper[4814]: I0227 16:56:29.110234 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk"] Feb 27 16:56:29 crc kubenswrapper[4814]: I0227 16:56:29.282183 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" event={"ID":"223a940a-b149-4b9a-a328-411a730cf5fb","Type":"ContainerStarted","Data":"caacab6bd531285b5a5877e795956fabbe89d39155e4a708c021b7bef0891a43"} Feb 27 16:56:30 crc kubenswrapper[4814]: I0227 16:56:30.310604 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" event={"ID":"223a940a-b149-4b9a-a328-411a730cf5fb","Type":"ContainerStarted","Data":"02df7747740a901cd0a55a2748b356a1c2a76b6284fe4f7af32ca4df2c64e777"} Feb 27 16:56:30 crc kubenswrapper[4814]: I0227 16:56:30.344242 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" podStartSLOduration=1.923441313 podStartE2EDuration="2.344210625s" podCreationTimestamp="2026-02-27 16:56:28 +0000 UTC" firstStartedPulling="2026-02-27 16:56:29.114092725 +0000 UTC m=+2001.566717555" lastFinishedPulling="2026-02-27 16:56:29.534861997 +0000 UTC m=+2001.987486867" observedRunningTime="2026-02-27 16:56:30.334650563 +0000 UTC m=+2002.787275423" watchObservedRunningTime="2026-02-27 16:56:30.344210625 +0000 UTC m=+2002.796835505" Feb 27 16:56:30 crc kubenswrapper[4814]: I0227 16:56:30.510536 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bbf0f83-c41f-427e-b2b4-6dffd46c5d09" path="/var/lib/kubelet/pods/6bbf0f83-c41f-427e-b2b4-6dffd46c5d09/volumes" Feb 27 16:56:31 crc kubenswrapper[4814]: I0227 16:56:31.487223 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:56:31 crc kubenswrapper[4814]: E0227 16:56:31.488071 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:56:38 crc kubenswrapper[4814]: I0227 16:56:38.403801 4814 generic.go:334] "Generic (PLEG): container finished" podID="223a940a-b149-4b9a-a328-411a730cf5fb" containerID="02df7747740a901cd0a55a2748b356a1c2a76b6284fe4f7af32ca4df2c64e777" exitCode=0 Feb 27 16:56:38 crc kubenswrapper[4814]: I0227 16:56:38.403987 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" event={"ID":"223a940a-b149-4b9a-a328-411a730cf5fb","Type":"ContainerDied","Data":"02df7747740a901cd0a55a2748b356a1c2a76b6284fe4f7af32ca4df2c64e777"} Feb 27 16:56:39 crc kubenswrapper[4814]: I0227 16:56:39.943414 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" Feb 27 16:56:39 crc kubenswrapper[4814]: I0227 16:56:39.991306 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/223a940a-b149-4b9a-a328-411a730cf5fb-ssh-key-openstack-edpm-ipam\") pod \"223a940a-b149-4b9a-a328-411a730cf5fb\" (UID: \"223a940a-b149-4b9a-a328-411a730cf5fb\") " Feb 27 16:56:39 crc kubenswrapper[4814]: I0227 16:56:39.991786 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/223a940a-b149-4b9a-a328-411a730cf5fb-inventory\") pod \"223a940a-b149-4b9a-a328-411a730cf5fb\" (UID: \"223a940a-b149-4b9a-a328-411a730cf5fb\") " Feb 27 16:56:39 crc kubenswrapper[4814]: I0227 16:56:39.991887 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f6r4\" (UniqueName: \"kubernetes.io/projected/223a940a-b149-4b9a-a328-411a730cf5fb-kube-api-access-9f6r4\") pod \"223a940a-b149-4b9a-a328-411a730cf5fb\" (UID: \"223a940a-b149-4b9a-a328-411a730cf5fb\") " Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.000846 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/223a940a-b149-4b9a-a328-411a730cf5fb-kube-api-access-9f6r4" (OuterVolumeSpecName: "kube-api-access-9f6r4") pod "223a940a-b149-4b9a-a328-411a730cf5fb" (UID: "223a940a-b149-4b9a-a328-411a730cf5fb"). InnerVolumeSpecName "kube-api-access-9f6r4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.030089 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/223a940a-b149-4b9a-a328-411a730cf5fb-inventory" (OuterVolumeSpecName: "inventory") pod "223a940a-b149-4b9a-a328-411a730cf5fb" (UID: "223a940a-b149-4b9a-a328-411a730cf5fb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.032428 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/223a940a-b149-4b9a-a328-411a730cf5fb-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "223a940a-b149-4b9a-a328-411a730cf5fb" (UID: "223a940a-b149-4b9a-a328-411a730cf5fb"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.094937 4814 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/223a940a-b149-4b9a-a328-411a730cf5fb-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.094979 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f6r4\" (UniqueName: \"kubernetes.io/projected/223a940a-b149-4b9a-a328-411a730cf5fb-kube-api-access-9f6r4\") on node \"crc\" DevicePath \"\"" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.094999 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/223a940a-b149-4b9a-a328-411a730cf5fb-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.436125 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" event={"ID":"223a940a-b149-4b9a-a328-411a730cf5fb","Type":"ContainerDied","Data":"caacab6bd531285b5a5877e795956fabbe89d39155e4a708c021b7bef0891a43"} Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.436192 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caacab6bd531285b5a5877e795956fabbe89d39155e4a708c021b7bef0891a43" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.436347 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s5rjk" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.647185 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9"] Feb 27 16:56:40 crc kubenswrapper[4814]: E0227 16:56:40.647853 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="223a940a-b149-4b9a-a328-411a730cf5fb" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.647883 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="223a940a-b149-4b9a-a328-411a730cf5fb" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.648227 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="223a940a-b149-4b9a-a328-411a730cf5fb" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.649422 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.652472 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.652493 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.653613 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.653850 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.666810 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9"] Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.705630 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9\" (UID: \"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.705839 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74vsz\" (UniqueName: \"kubernetes.io/projected/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-kube-api-access-74vsz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9\" (UID: \"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.705987 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9\" (UID: \"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.807862 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74vsz\" (UniqueName: \"kubernetes.io/projected/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-kube-api-access-74vsz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9\" (UID: \"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.807945 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9\" (UID: \"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.808041 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9\" (UID: \"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.814050 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9\" (UID: \"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.814448 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9\" (UID: \"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.830966 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74vsz\" (UniqueName: \"kubernetes.io/projected/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-kube-api-access-74vsz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9\" (UID: \"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" Feb 27 16:56:40 crc kubenswrapper[4814]: I0227 16:56:40.987128 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" Feb 27 16:56:41 crc kubenswrapper[4814]: I0227 16:56:41.507120 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9"] Feb 27 16:56:42 crc kubenswrapper[4814]: I0227 16:56:42.458248 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" event={"ID":"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3","Type":"ContainerStarted","Data":"e043c1bb98d9ada6db1d510268bddd314a80c68db09f8159eeb973b41109ab71"} Feb 27 16:56:42 crc kubenswrapper[4814]: I0227 16:56:42.458792 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" event={"ID":"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3","Type":"ContainerStarted","Data":"19f27d8641ede6b5b252e5427ef57a624ad0a72044bf4f6dbee0d64132de9096"} Feb 27 16:56:42 crc kubenswrapper[4814]: I0227 16:56:42.500976 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" podStartSLOduration=2.065867451 podStartE2EDuration="2.500947811s" podCreationTimestamp="2026-02-27 16:56:40 +0000 UTC" firstStartedPulling="2026-02-27 16:56:41.516552881 +0000 UTC m=+2013.969177721" lastFinishedPulling="2026-02-27 16:56:41.951633241 +0000 UTC m=+2014.404258081" observedRunningTime="2026-02-27 16:56:42.483234239 +0000 UTC m=+2014.935859099" watchObservedRunningTime="2026-02-27 16:56:42.500947811 +0000 UTC m=+2014.953572681" Feb 27 16:56:46 crc kubenswrapper[4814]: I0227 16:56:46.488982 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:56:46 crc kubenswrapper[4814]: E0227 16:56:46.489867 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:56:52 crc kubenswrapper[4814]: I0227 16:56:52.582235 4814 generic.go:334] "Generic (PLEG): container finished" podID="4cd3437d-9ecc-44b6-b9e7-7aaed24147a3" containerID="e043c1bb98d9ada6db1d510268bddd314a80c68db09f8159eeb973b41109ab71" exitCode=0 Feb 27 16:56:52 crc kubenswrapper[4814]: I0227 16:56:52.582422 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" event={"ID":"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3","Type":"ContainerDied","Data":"e043c1bb98d9ada6db1d510268bddd314a80c68db09f8159eeb973b41109ab71"} Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.105678 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.220409 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74vsz\" (UniqueName: \"kubernetes.io/projected/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-kube-api-access-74vsz\") pod \"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3\" (UID: \"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3\") " Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.220487 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-inventory\") pod \"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3\" (UID: \"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3\") " Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.220538 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-ssh-key-openstack-edpm-ipam\") pod \"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3\" (UID: \"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3\") " Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.226045 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-kube-api-access-74vsz" (OuterVolumeSpecName: "kube-api-access-74vsz") pod "4cd3437d-9ecc-44b6-b9e7-7aaed24147a3" (UID: "4cd3437d-9ecc-44b6-b9e7-7aaed24147a3"). InnerVolumeSpecName "kube-api-access-74vsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.260446 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-inventory" (OuterVolumeSpecName: "inventory") pod "4cd3437d-9ecc-44b6-b9e7-7aaed24147a3" (UID: "4cd3437d-9ecc-44b6-b9e7-7aaed24147a3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.261625 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4cd3437d-9ecc-44b6-b9e7-7aaed24147a3" (UID: "4cd3437d-9ecc-44b6-b9e7-7aaed24147a3"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.323983 4814 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.324043 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.324073 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74vsz\" (UniqueName: \"kubernetes.io/projected/4cd3437d-9ecc-44b6-b9e7-7aaed24147a3-kube-api-access-74vsz\") on node \"crc\" DevicePath \"\"" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.609669 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" event={"ID":"4cd3437d-9ecc-44b6-b9e7-7aaed24147a3","Type":"ContainerDied","Data":"19f27d8641ede6b5b252e5427ef57a624ad0a72044bf4f6dbee0d64132de9096"} Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.609710 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19f27d8641ede6b5b252e5427ef57a624ad0a72044bf4f6dbee0d64132de9096" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.609715 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.710987 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56"] Feb 27 16:56:54 crc kubenswrapper[4814]: E0227 16:56:54.711440 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cd3437d-9ecc-44b6-b9e7-7aaed24147a3" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.711462 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cd3437d-9ecc-44b6-b9e7-7aaed24147a3" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.711670 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cd3437d-9ecc-44b6-b9e7-7aaed24147a3" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.712372 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.715939 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.716207 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.716331 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.716398 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.716726 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.717089 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.717091 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.717942 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.738894 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56"] Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.835591 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.835653 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.835680 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.837099 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.837213 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.837246 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrkzd\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-kube-api-access-qrkzd\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.837344 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.837402 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.837429 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.837522 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.837573 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.837604 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.837630 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.837765 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.939597 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.939678 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.939719 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.939784 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.939862 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.939908 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrkzd\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-kube-api-access-qrkzd\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.939979 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.940044 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.940084 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.940161 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.940228 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.940309 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.940365 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.940472 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.945591 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.946615 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.946703 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.946824 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.947852 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.947906 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.948638 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.949619 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.950677 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.950884 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.951561 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.951663 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.958499 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:54 crc kubenswrapper[4814]: I0227 16:56:54.965856 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrkzd\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-kube-api-access-qrkzd\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-btx56\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:55 crc kubenswrapper[4814]: I0227 16:56:55.039115 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:56:55 crc kubenswrapper[4814]: I0227 16:56:55.702777 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56"] Feb 27 16:56:56 crc kubenswrapper[4814]: I0227 16:56:56.462018 4814 scope.go:117] "RemoveContainer" containerID="c7592a56dad2fc0610cf175f91deea1aaf285a476965cc1103a83e5c84d3d6cf" Feb 27 16:56:56 crc kubenswrapper[4814]: I0227 16:56:56.530168 4814 scope.go:117] "RemoveContainer" containerID="a26226c1f8ce87ab249147fa48629d5c0bb6dc47990436b44cfe34d29a5d9ff1" Feb 27 16:56:56 crc kubenswrapper[4814]: I0227 16:56:56.634823 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" event={"ID":"d9f3ab2c-0157-4a91-a329-c09648662cce","Type":"ContainerStarted","Data":"173a50b9d4699ebeebd07ef64957f64ba55b273e82ff079c7f82dbe5de491df2"} Feb 27 16:56:57 crc kubenswrapper[4814]: I0227 16:56:57.645531 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" event={"ID":"d9f3ab2c-0157-4a91-a329-c09648662cce","Type":"ContainerStarted","Data":"7a5cff9d48741286d90264ccb874e2923196d9be376027fd39fa62606f6d9b14"} Feb 27 16:56:57 crc kubenswrapper[4814]: I0227 16:56:57.672218 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" podStartSLOduration=3.018653541 podStartE2EDuration="3.672199238s" podCreationTimestamp="2026-02-27 16:56:54 +0000 UTC" firstStartedPulling="2026-02-27 16:56:55.709755923 +0000 UTC m=+2028.162380753" lastFinishedPulling="2026-02-27 16:56:56.36330162 +0000 UTC m=+2028.815926450" observedRunningTime="2026-02-27 16:56:57.667728432 +0000 UTC m=+2030.120353262" watchObservedRunningTime="2026-02-27 16:56:57.672199238 +0000 UTC m=+2030.124824068" Feb 27 16:57:00 crc kubenswrapper[4814]: I0227 16:57:00.488696 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:57:00 crc kubenswrapper[4814]: E0227 16:57:00.489727 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:57:14 crc kubenswrapper[4814]: I0227 16:57:14.487898 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:57:14 crc kubenswrapper[4814]: E0227 16:57:14.488996 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:57:28 crc kubenswrapper[4814]: I0227 16:57:28.496911 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:57:28 crc kubenswrapper[4814]: E0227 16:57:28.497715 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:57:39 crc kubenswrapper[4814]: I0227 16:57:39.125467 4814 generic.go:334] "Generic (PLEG): container finished" podID="d9f3ab2c-0157-4a91-a329-c09648662cce" containerID="7a5cff9d48741286d90264ccb874e2923196d9be376027fd39fa62606f6d9b14" exitCode=0 Feb 27 16:57:39 crc kubenswrapper[4814]: I0227 16:57:39.125557 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" event={"ID":"d9f3ab2c-0157-4a91-a329-c09648662cce","Type":"ContainerDied","Data":"7a5cff9d48741286d90264ccb874e2923196d9be376027fd39fa62606f6d9b14"} Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.489032 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:57:40 crc kubenswrapper[4814]: E0227 16:57:40.490018 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.690238 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.834237 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"d9f3ab2c-0157-4a91-a329-c09648662cce\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.834359 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-ssh-key-openstack-edpm-ipam\") pod \"d9f3ab2c-0157-4a91-a329-c09648662cce\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.834405 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"d9f3ab2c-0157-4a91-a329-c09648662cce\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.834439 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-libvirt-combined-ca-bundle\") pod \"d9f3ab2c-0157-4a91-a329-c09648662cce\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.834467 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-inventory\") pod \"d9f3ab2c-0157-4a91-a329-c09648662cce\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.834580 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-repo-setup-combined-ca-bundle\") pod \"d9f3ab2c-0157-4a91-a329-c09648662cce\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.835588 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-neutron-metadata-combined-ca-bundle\") pod \"d9f3ab2c-0157-4a91-a329-c09648662cce\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.835666 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-telemetry-combined-ca-bundle\") pod \"d9f3ab2c-0157-4a91-a329-c09648662cce\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.835691 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-ovn-combined-ca-bundle\") pod \"d9f3ab2c-0157-4a91-a329-c09648662cce\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.835737 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrkzd\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-kube-api-access-qrkzd\") pod \"d9f3ab2c-0157-4a91-a329-c09648662cce\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.835828 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-bootstrap-combined-ca-bundle\") pod \"d9f3ab2c-0157-4a91-a329-c09648662cce\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.835860 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-nova-combined-ca-bundle\") pod \"d9f3ab2c-0157-4a91-a329-c09648662cce\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.835911 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-ovn-default-certs-0\") pod \"d9f3ab2c-0157-4a91-a329-c09648662cce\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.835951 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"d9f3ab2c-0157-4a91-a329-c09648662cce\" (UID: \"d9f3ab2c-0157-4a91-a329-c09648662cce\") " Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.843336 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d9f3ab2c-0157-4a91-a329-c09648662cce" (UID: "d9f3ab2c-0157-4a91-a329-c09648662cce"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.843441 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d9f3ab2c-0157-4a91-a329-c09648662cce" (UID: "d9f3ab2c-0157-4a91-a329-c09648662cce"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.843776 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d9f3ab2c-0157-4a91-a329-c09648662cce" (UID: "d9f3ab2c-0157-4a91-a329-c09648662cce"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.843891 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "d9f3ab2c-0157-4a91-a329-c09648662cce" (UID: "d9f3ab2c-0157-4a91-a329-c09648662cce"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.844267 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d9f3ab2c-0157-4a91-a329-c09648662cce" (UID: "d9f3ab2c-0157-4a91-a329-c09648662cce"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.844448 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "d9f3ab2c-0157-4a91-a329-c09648662cce" (UID: "d9f3ab2c-0157-4a91-a329-c09648662cce"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.845053 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "d9f3ab2c-0157-4a91-a329-c09648662cce" (UID: "d9f3ab2c-0157-4a91-a329-c09648662cce"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.845677 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d9f3ab2c-0157-4a91-a329-c09648662cce" (UID: "d9f3ab2c-0157-4a91-a329-c09648662cce"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.847672 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d9f3ab2c-0157-4a91-a329-c09648662cce" (UID: "d9f3ab2c-0157-4a91-a329-c09648662cce"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.847870 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-kube-api-access-qrkzd" (OuterVolumeSpecName: "kube-api-access-qrkzd") pod "d9f3ab2c-0157-4a91-a329-c09648662cce" (UID: "d9f3ab2c-0157-4a91-a329-c09648662cce"). InnerVolumeSpecName "kube-api-access-qrkzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.850512 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "d9f3ab2c-0157-4a91-a329-c09648662cce" (UID: "d9f3ab2c-0157-4a91-a329-c09648662cce"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.853479 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d9f3ab2c-0157-4a91-a329-c09648662cce" (UID: "d9f3ab2c-0157-4a91-a329-c09648662cce"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.884211 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d9f3ab2c-0157-4a91-a329-c09648662cce" (UID: "d9f3ab2c-0157-4a91-a329-c09648662cce"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.885910 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-inventory" (OuterVolumeSpecName: "inventory") pod "d9f3ab2c-0157-4a91-a329-c09648662cce" (UID: "d9f3ab2c-0157-4a91-a329-c09648662cce"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.939380 4814 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.939461 4814 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.939472 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrkzd\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-kube-api-access-qrkzd\") on node \"crc\" DevicePath \"\"" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.939486 4814 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.939497 4814 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.939508 4814 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.939521 4814 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.939541 4814 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.939558 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.939568 4814 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d9f3ab2c-0157-4a91-a329-c09648662cce-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.939578 4814 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.939589 4814 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.939599 4814 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:57:40 crc kubenswrapper[4814]: I0227 16:57:40.939611 4814 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f3ab2c-0157-4a91-a329-c09648662cce-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.155575 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" event={"ID":"d9f3ab2c-0157-4a91-a329-c09648662cce","Type":"ContainerDied","Data":"173a50b9d4699ebeebd07ef64957f64ba55b273e82ff079c7f82dbe5de491df2"} Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.155612 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="173a50b9d4699ebeebd07ef64957f64ba55b273e82ff079c7f82dbe5de491df2" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.155693 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-btx56" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.285814 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd"] Feb 27 16:57:41 crc kubenswrapper[4814]: E0227 16:57:41.286552 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9f3ab2c-0157-4a91-a329-c09648662cce" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.286577 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9f3ab2c-0157-4a91-a329-c09648662cce" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.287168 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9f3ab2c-0157-4a91-a329-c09648662cce" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.288208 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.290585 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.292186 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.292790 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.293730 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.295620 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.299807 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd"] Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.450109 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-58prd\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.450176 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx6gh\" (UniqueName: \"kubernetes.io/projected/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-kube-api-access-kx6gh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-58prd\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.450196 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-58prd\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.450315 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-58prd\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.450358 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-58prd\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.552161 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-58prd\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.552224 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-58prd\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.552280 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-58prd\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.552322 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx6gh\" (UniqueName: \"kubernetes.io/projected/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-kube-api-access-kx6gh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-58prd\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.552343 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-58prd\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.553210 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-58prd\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.557362 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-58prd\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.559987 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-58prd\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.564818 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-58prd\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.572640 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx6gh\" (UniqueName: \"kubernetes.io/projected/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-kube-api-access-kx6gh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-58prd\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:41 crc kubenswrapper[4814]: I0227 16:57:41.621404 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:57:42 crc kubenswrapper[4814]: I0227 16:57:42.290892 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd"] Feb 27 16:57:43 crc kubenswrapper[4814]: I0227 16:57:43.181201 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" event={"ID":"245bd73e-7f47-4c71-9e26-c2b74c8b58e2","Type":"ContainerStarted","Data":"714d5810d7a1c0bb08c6da71a7ae5b8b9ad554b7b69f5cc05ba3dc31079d3a24"} Feb 27 16:57:44 crc kubenswrapper[4814]: I0227 16:57:44.198684 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" event={"ID":"245bd73e-7f47-4c71-9e26-c2b74c8b58e2","Type":"ContainerStarted","Data":"92cf27631dfa4e8f72c6a634220d9611ef35c3b098a0d1567af3b35f86a038be"} Feb 27 16:57:44 crc kubenswrapper[4814]: I0227 16:57:44.236769 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" podStartSLOduration=2.657919816 podStartE2EDuration="3.236743651s" podCreationTimestamp="2026-02-27 16:57:41 +0000 UTC" firstStartedPulling="2026-02-27 16:57:42.304469805 +0000 UTC m=+2074.757094635" lastFinishedPulling="2026-02-27 16:57:42.88329362 +0000 UTC m=+2075.335918470" observedRunningTime="2026-02-27 16:57:44.22765424 +0000 UTC m=+2076.680279120" watchObservedRunningTime="2026-02-27 16:57:44.236743651 +0000 UTC m=+2076.689368491" Feb 27 16:57:55 crc kubenswrapper[4814]: I0227 16:57:55.488561 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 16:57:56 crc kubenswrapper[4814]: I0227 16:57:56.336180 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"eb9a72219e9b318e493f16a1ab078d7373596a6f5f4b8c4c5a593b8ea470dc7c"} Feb 27 16:58:00 crc kubenswrapper[4814]: I0227 16:58:00.151615 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536858-z4w2z"] Feb 27 16:58:00 crc kubenswrapper[4814]: I0227 16:58:00.154902 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536858-z4w2z" Feb 27 16:58:00 crc kubenswrapper[4814]: I0227 16:58:00.159180 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 16:58:00 crc kubenswrapper[4814]: I0227 16:58:00.159433 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 16:58:00 crc kubenswrapper[4814]: I0227 16:58:00.159847 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 16:58:00 crc kubenswrapper[4814]: I0227 16:58:00.165097 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536858-z4w2z"] Feb 27 16:58:00 crc kubenswrapper[4814]: I0227 16:58:00.231313 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwm6w\" (UniqueName: \"kubernetes.io/projected/0060b006-c2cf-4a60-87b3-5c9f644d0ad9-kube-api-access-kwm6w\") pod \"auto-csr-approver-29536858-z4w2z\" (UID: \"0060b006-c2cf-4a60-87b3-5c9f644d0ad9\") " pod="openshift-infra/auto-csr-approver-29536858-z4w2z" Feb 27 16:58:00 crc kubenswrapper[4814]: I0227 16:58:00.341062 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwm6w\" (UniqueName: \"kubernetes.io/projected/0060b006-c2cf-4a60-87b3-5c9f644d0ad9-kube-api-access-kwm6w\") pod \"auto-csr-approver-29536858-z4w2z\" (UID: \"0060b006-c2cf-4a60-87b3-5c9f644d0ad9\") " pod="openshift-infra/auto-csr-approver-29536858-z4w2z" Feb 27 16:58:00 crc kubenswrapper[4814]: I0227 16:58:00.360468 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwm6w\" (UniqueName: \"kubernetes.io/projected/0060b006-c2cf-4a60-87b3-5c9f644d0ad9-kube-api-access-kwm6w\") pod \"auto-csr-approver-29536858-z4w2z\" (UID: \"0060b006-c2cf-4a60-87b3-5c9f644d0ad9\") " pod="openshift-infra/auto-csr-approver-29536858-z4w2z" Feb 27 16:58:00 crc kubenswrapper[4814]: I0227 16:58:00.488991 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536858-z4w2z" Feb 27 16:58:00 crc kubenswrapper[4814]: I0227 16:58:00.980129 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536858-z4w2z"] Feb 27 16:58:01 crc kubenswrapper[4814]: I0227 16:58:01.387510 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536858-z4w2z" event={"ID":"0060b006-c2cf-4a60-87b3-5c9f644d0ad9","Type":"ContainerStarted","Data":"31b489eaaf7249dd9fcd7e50a3d0575fd6399c30351c59af4cc56b86c3d7e030"} Feb 27 16:58:02 crc kubenswrapper[4814]: I0227 16:58:02.396993 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536858-z4w2z" event={"ID":"0060b006-c2cf-4a60-87b3-5c9f644d0ad9","Type":"ContainerStarted","Data":"adbef7afb182a7c61ac9f9b627652bbae787ceee9e64bb40aa473679d4df7cf4"} Feb 27 16:58:02 crc kubenswrapper[4814]: I0227 16:58:02.417665 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536858-z4w2z" podStartSLOduration=1.525934102 podStartE2EDuration="2.41764548s" podCreationTimestamp="2026-02-27 16:58:00 +0000 UTC" firstStartedPulling="2026-02-27 16:58:00.979600539 +0000 UTC m=+2093.432225369" lastFinishedPulling="2026-02-27 16:58:01.871311917 +0000 UTC m=+2094.323936747" observedRunningTime="2026-02-27 16:58:02.411704787 +0000 UTC m=+2094.864329637" watchObservedRunningTime="2026-02-27 16:58:02.41764548 +0000 UTC m=+2094.870270310" Feb 27 16:58:03 crc kubenswrapper[4814]: I0227 16:58:03.405685 4814 generic.go:334] "Generic (PLEG): container finished" podID="0060b006-c2cf-4a60-87b3-5c9f644d0ad9" containerID="adbef7afb182a7c61ac9f9b627652bbae787ceee9e64bb40aa473679d4df7cf4" exitCode=0 Feb 27 16:58:03 crc kubenswrapper[4814]: I0227 16:58:03.405764 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536858-z4w2z" event={"ID":"0060b006-c2cf-4a60-87b3-5c9f644d0ad9","Type":"ContainerDied","Data":"adbef7afb182a7c61ac9f9b627652bbae787ceee9e64bb40aa473679d4df7cf4"} Feb 27 16:58:04 crc kubenswrapper[4814]: I0227 16:58:04.770790 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536858-z4w2z" Feb 27 16:58:04 crc kubenswrapper[4814]: I0227 16:58:04.836910 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwm6w\" (UniqueName: \"kubernetes.io/projected/0060b006-c2cf-4a60-87b3-5c9f644d0ad9-kube-api-access-kwm6w\") pod \"0060b006-c2cf-4a60-87b3-5c9f644d0ad9\" (UID: \"0060b006-c2cf-4a60-87b3-5c9f644d0ad9\") " Feb 27 16:58:04 crc kubenswrapper[4814]: I0227 16:58:04.844682 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0060b006-c2cf-4a60-87b3-5c9f644d0ad9-kube-api-access-kwm6w" (OuterVolumeSpecName: "kube-api-access-kwm6w") pod "0060b006-c2cf-4a60-87b3-5c9f644d0ad9" (UID: "0060b006-c2cf-4a60-87b3-5c9f644d0ad9"). InnerVolumeSpecName "kube-api-access-kwm6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:58:04 crc kubenswrapper[4814]: I0227 16:58:04.943132 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwm6w\" (UniqueName: \"kubernetes.io/projected/0060b006-c2cf-4a60-87b3-5c9f644d0ad9-kube-api-access-kwm6w\") on node \"crc\" DevicePath \"\"" Feb 27 16:58:05 crc kubenswrapper[4814]: I0227 16:58:05.436955 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536858-z4w2z" event={"ID":"0060b006-c2cf-4a60-87b3-5c9f644d0ad9","Type":"ContainerDied","Data":"31b489eaaf7249dd9fcd7e50a3d0575fd6399c30351c59af4cc56b86c3d7e030"} Feb 27 16:58:05 crc kubenswrapper[4814]: I0227 16:58:05.437350 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31b489eaaf7249dd9fcd7e50a3d0575fd6399c30351c59af4cc56b86c3d7e030" Feb 27 16:58:05 crc kubenswrapper[4814]: I0227 16:58:05.437021 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536858-z4w2z" Feb 27 16:58:05 crc kubenswrapper[4814]: I0227 16:58:05.500346 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536852-8vpsm"] Feb 27 16:58:05 crc kubenswrapper[4814]: I0227 16:58:05.508592 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536852-8vpsm"] Feb 27 16:58:06 crc kubenswrapper[4814]: I0227 16:58:06.507640 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93b95ff2-824a-43ff-9110-ef73f1fcc34c" path="/var/lib/kubelet/pods/93b95ff2-824a-43ff-9110-ef73f1fcc34c/volumes" Feb 27 16:58:53 crc kubenswrapper[4814]: I0227 16:58:53.994224 4814 generic.go:334] "Generic (PLEG): container finished" podID="245bd73e-7f47-4c71-9e26-c2b74c8b58e2" containerID="92cf27631dfa4e8f72c6a634220d9611ef35c3b098a0d1567af3b35f86a038be" exitCode=0 Feb 27 16:58:53 crc kubenswrapper[4814]: I0227 16:58:53.994576 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" event={"ID":"245bd73e-7f47-4c71-9e26-c2b74c8b58e2","Type":"ContainerDied","Data":"92cf27631dfa4e8f72c6a634220d9611ef35c3b098a0d1567af3b35f86a038be"} Feb 27 16:58:55 crc kubenswrapper[4814]: I0227 16:58:55.455593 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:58:55 crc kubenswrapper[4814]: I0227 16:58:55.583015 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ovn-combined-ca-bundle\") pod \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " Feb 27 16:58:55 crc kubenswrapper[4814]: I0227 16:58:55.583123 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ovncontroller-config-0\") pod \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " Feb 27 16:58:55 crc kubenswrapper[4814]: I0227 16:58:55.583161 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ssh-key-openstack-edpm-ipam\") pod \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " Feb 27 16:58:55 crc kubenswrapper[4814]: I0227 16:58:55.583201 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx6gh\" (UniqueName: \"kubernetes.io/projected/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-kube-api-access-kx6gh\") pod \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " Feb 27 16:58:55 crc kubenswrapper[4814]: I0227 16:58:55.583240 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-inventory\") pod \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\" (UID: \"245bd73e-7f47-4c71-9e26-c2b74c8b58e2\") " Feb 27 16:58:55 crc kubenswrapper[4814]: I0227 16:58:55.591021 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-kube-api-access-kx6gh" (OuterVolumeSpecName: "kube-api-access-kx6gh") pod "245bd73e-7f47-4c71-9e26-c2b74c8b58e2" (UID: "245bd73e-7f47-4c71-9e26-c2b74c8b58e2"). InnerVolumeSpecName "kube-api-access-kx6gh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:58:55 crc kubenswrapper[4814]: I0227 16:58:55.591194 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "245bd73e-7f47-4c71-9e26-c2b74c8b58e2" (UID: "245bd73e-7f47-4c71-9e26-c2b74c8b58e2"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:58:55 crc kubenswrapper[4814]: I0227 16:58:55.625604 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "245bd73e-7f47-4c71-9e26-c2b74c8b58e2" (UID: "245bd73e-7f47-4c71-9e26-c2b74c8b58e2"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 16:58:55 crc kubenswrapper[4814]: I0227 16:58:55.631722 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "245bd73e-7f47-4c71-9e26-c2b74c8b58e2" (UID: "245bd73e-7f47-4c71-9e26-c2b74c8b58e2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:58:55 crc kubenswrapper[4814]: I0227 16:58:55.649477 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-inventory" (OuterVolumeSpecName: "inventory") pod "245bd73e-7f47-4c71-9e26-c2b74c8b58e2" (UID: "245bd73e-7f47-4c71-9e26-c2b74c8b58e2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:58:55 crc kubenswrapper[4814]: I0227 16:58:55.686735 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 16:58:55 crc kubenswrapper[4814]: I0227 16:58:55.686769 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx6gh\" (UniqueName: \"kubernetes.io/projected/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-kube-api-access-kx6gh\") on node \"crc\" DevicePath \"\"" Feb 27 16:58:55 crc kubenswrapper[4814]: I0227 16:58:55.686779 4814 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 16:58:55 crc kubenswrapper[4814]: I0227 16:58:55.686789 4814 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:58:55 crc kubenswrapper[4814]: I0227 16:58:55.686798 4814 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/245bd73e-7f47-4c71-9e26-c2b74c8b58e2-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.019076 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" event={"ID":"245bd73e-7f47-4c71-9e26-c2b74c8b58e2","Type":"ContainerDied","Data":"714d5810d7a1c0bb08c6da71a7ae5b8b9ad554b7b69f5cc05ba3dc31079d3a24"} Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.019135 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="714d5810d7a1c0bb08c6da71a7ae5b8b9ad554b7b69f5cc05ba3dc31079d3a24" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.019177 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-58prd" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.142017 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88"] Feb 27 16:58:56 crc kubenswrapper[4814]: E0227 16:58:56.142554 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0060b006-c2cf-4a60-87b3-5c9f644d0ad9" containerName="oc" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.142580 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="0060b006-c2cf-4a60-87b3-5c9f644d0ad9" containerName="oc" Feb 27 16:58:56 crc kubenswrapper[4814]: E0227 16:58:56.142623 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="245bd73e-7f47-4c71-9e26-c2b74c8b58e2" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.142633 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="245bd73e-7f47-4c71-9e26-c2b74c8b58e2" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.142798 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="0060b006-c2cf-4a60-87b3-5c9f644d0ad9" containerName="oc" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.142826 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="245bd73e-7f47-4c71-9e26-c2b74c8b58e2" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.143469 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.149490 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.154621 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.154776 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.154881 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.154987 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.155054 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.170690 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88"] Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.302487 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkgk6\" (UniqueName: \"kubernetes.io/projected/cf70174d-7319-418e-8545-83db61a7eb7c-kube-api-access-pkgk6\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.302548 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.302587 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.302609 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.302834 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.302968 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.404690 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.404782 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkgk6\" (UniqueName: \"kubernetes.io/projected/cf70174d-7319-418e-8545-83db61a7eb7c-kube-api-access-pkgk6\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.404828 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.404861 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.404897 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.404963 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.412311 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.414882 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.415026 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.415340 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.417999 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.432411 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkgk6\" (UniqueName: \"kubernetes.io/projected/cf70174d-7319-418e-8545-83db61a7eb7c-kube-api-access-pkgk6\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.464533 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:58:56 crc kubenswrapper[4814]: I0227 16:58:56.652158 4814 scope.go:117] "RemoveContainer" containerID="194d3a56e8652d3331106c1113dd11dda49c95bc664e6972c88716ae024dc4e7" Feb 27 16:58:57 crc kubenswrapper[4814]: I0227 16:58:57.034182 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88"] Feb 27 16:58:57 crc kubenswrapper[4814]: I0227 16:58:57.043858 4814 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 16:58:58 crc kubenswrapper[4814]: I0227 16:58:58.037705 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" event={"ID":"cf70174d-7319-418e-8545-83db61a7eb7c","Type":"ContainerStarted","Data":"18d10b2dc0d47e66434ca4914307de01a647006be6f3c1c0e2492fd3ce62d321"} Feb 27 16:58:58 crc kubenswrapper[4814]: I0227 16:58:58.038086 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" event={"ID":"cf70174d-7319-418e-8545-83db61a7eb7c","Type":"ContainerStarted","Data":"66d594822bde897c5cb075c5611d94898b13ece518e899d539c9852f5d5101a7"} Feb 27 16:58:58 crc kubenswrapper[4814]: I0227 16:58:58.065538 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" podStartSLOduration=1.598549574 podStartE2EDuration="2.065519339s" podCreationTimestamp="2026-02-27 16:58:56 +0000 UTC" firstStartedPulling="2026-02-27 16:58:57.043655807 +0000 UTC m=+2149.496280627" lastFinishedPulling="2026-02-27 16:58:57.510625532 +0000 UTC m=+2149.963250392" observedRunningTime="2026-02-27 16:58:58.060156124 +0000 UTC m=+2150.512780954" watchObservedRunningTime="2026-02-27 16:58:58.065519339 +0000 UTC m=+2150.518144169" Feb 27 16:59:03 crc kubenswrapper[4814]: I0227 16:59:03.644597 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r9q8w"] Feb 27 16:59:03 crc kubenswrapper[4814]: I0227 16:59:03.646707 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:03 crc kubenswrapper[4814]: I0227 16:59:03.663787 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r9q8w"] Feb 27 16:59:03 crc kubenswrapper[4814]: I0227 16:59:03.805034 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0b4e818-ea6c-4450-be75-44c769aeb159-catalog-content\") pod \"redhat-operators-r9q8w\" (UID: \"a0b4e818-ea6c-4450-be75-44c769aeb159\") " pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:03 crc kubenswrapper[4814]: I0227 16:59:03.805106 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh96j\" (UniqueName: \"kubernetes.io/projected/a0b4e818-ea6c-4450-be75-44c769aeb159-kube-api-access-fh96j\") pod \"redhat-operators-r9q8w\" (UID: \"a0b4e818-ea6c-4450-be75-44c769aeb159\") " pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:03 crc kubenswrapper[4814]: I0227 16:59:03.805136 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0b4e818-ea6c-4450-be75-44c769aeb159-utilities\") pod \"redhat-operators-r9q8w\" (UID: \"a0b4e818-ea6c-4450-be75-44c769aeb159\") " pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:03 crc kubenswrapper[4814]: I0227 16:59:03.907400 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0b4e818-ea6c-4450-be75-44c769aeb159-catalog-content\") pod \"redhat-operators-r9q8w\" (UID: \"a0b4e818-ea6c-4450-be75-44c769aeb159\") " pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:03 crc kubenswrapper[4814]: I0227 16:59:03.907454 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh96j\" (UniqueName: \"kubernetes.io/projected/a0b4e818-ea6c-4450-be75-44c769aeb159-kube-api-access-fh96j\") pod \"redhat-operators-r9q8w\" (UID: \"a0b4e818-ea6c-4450-be75-44c769aeb159\") " pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:03 crc kubenswrapper[4814]: I0227 16:59:03.907485 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0b4e818-ea6c-4450-be75-44c769aeb159-utilities\") pod \"redhat-operators-r9q8w\" (UID: \"a0b4e818-ea6c-4450-be75-44c769aeb159\") " pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:03 crc kubenswrapper[4814]: I0227 16:59:03.907995 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0b4e818-ea6c-4450-be75-44c769aeb159-catalog-content\") pod \"redhat-operators-r9q8w\" (UID: \"a0b4e818-ea6c-4450-be75-44c769aeb159\") " pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:03 crc kubenswrapper[4814]: I0227 16:59:03.908002 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0b4e818-ea6c-4450-be75-44c769aeb159-utilities\") pod \"redhat-operators-r9q8w\" (UID: \"a0b4e818-ea6c-4450-be75-44c769aeb159\") " pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:03 crc kubenswrapper[4814]: I0227 16:59:03.928978 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh96j\" (UniqueName: \"kubernetes.io/projected/a0b4e818-ea6c-4450-be75-44c769aeb159-kube-api-access-fh96j\") pod \"redhat-operators-r9q8w\" (UID: \"a0b4e818-ea6c-4450-be75-44c769aeb159\") " pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:03 crc kubenswrapper[4814]: I0227 16:59:03.964774 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:04 crc kubenswrapper[4814]: I0227 16:59:04.415500 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r9q8w"] Feb 27 16:59:05 crc kubenswrapper[4814]: I0227 16:59:05.101582 4814 generic.go:334] "Generic (PLEG): container finished" podID="a0b4e818-ea6c-4450-be75-44c769aeb159" containerID="6a68c95c4aedca638d932bacacdc9074568a7bb79c0f52cd2b928cf3acf5dfac" exitCode=0 Feb 27 16:59:05 crc kubenswrapper[4814]: I0227 16:59:05.101698 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9q8w" event={"ID":"a0b4e818-ea6c-4450-be75-44c769aeb159","Type":"ContainerDied","Data":"6a68c95c4aedca638d932bacacdc9074568a7bb79c0f52cd2b928cf3acf5dfac"} Feb 27 16:59:05 crc kubenswrapper[4814]: I0227 16:59:05.101872 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9q8w" event={"ID":"a0b4e818-ea6c-4450-be75-44c769aeb159","Type":"ContainerStarted","Data":"9e8bc408074eaed3c9946d2e7531e0ba618b9b78ed6493ddf7e72f76e286d0ce"} Feb 27 16:59:06 crc kubenswrapper[4814]: I0227 16:59:06.117315 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9q8w" event={"ID":"a0b4e818-ea6c-4450-be75-44c769aeb159","Type":"ContainerStarted","Data":"43678db89c371aae86da7c8239fb05f4b3603821af79c149f414218c828d3f48"} Feb 27 16:59:07 crc kubenswrapper[4814]: I0227 16:59:07.130103 4814 generic.go:334] "Generic (PLEG): container finished" podID="a0b4e818-ea6c-4450-be75-44c769aeb159" containerID="43678db89c371aae86da7c8239fb05f4b3603821af79c149f414218c828d3f48" exitCode=0 Feb 27 16:59:07 crc kubenswrapper[4814]: I0227 16:59:07.130184 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9q8w" event={"ID":"a0b4e818-ea6c-4450-be75-44c769aeb159","Type":"ContainerDied","Data":"43678db89c371aae86da7c8239fb05f4b3603821af79c149f414218c828d3f48"} Feb 27 16:59:08 crc kubenswrapper[4814]: I0227 16:59:08.152632 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9q8w" event={"ID":"a0b4e818-ea6c-4450-be75-44c769aeb159","Type":"ContainerStarted","Data":"786b544f3b74635eda559cb5099a3b686bc38e1bfbd78c94a282086b49814f8b"} Feb 27 16:59:08 crc kubenswrapper[4814]: I0227 16:59:08.187719 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r9q8w" podStartSLOduration=2.645588558 podStartE2EDuration="5.187698006s" podCreationTimestamp="2026-02-27 16:59:03 +0000 UTC" firstStartedPulling="2026-02-27 16:59:05.104101544 +0000 UTC m=+2157.556726374" lastFinishedPulling="2026-02-27 16:59:07.646210982 +0000 UTC m=+2160.098835822" observedRunningTime="2026-02-27 16:59:08.180209304 +0000 UTC m=+2160.632834144" watchObservedRunningTime="2026-02-27 16:59:08.187698006 +0000 UTC m=+2160.640322846" Feb 27 16:59:13 crc kubenswrapper[4814]: I0227 16:59:13.965923 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:13 crc kubenswrapper[4814]: I0227 16:59:13.966501 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:15 crc kubenswrapper[4814]: I0227 16:59:15.033159 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-r9q8w" podUID="a0b4e818-ea6c-4450-be75-44c769aeb159" containerName="registry-server" probeResult="failure" output=< Feb 27 16:59:15 crc kubenswrapper[4814]: timeout: failed to connect service ":50051" within 1s Feb 27 16:59:15 crc kubenswrapper[4814]: > Feb 27 16:59:24 crc kubenswrapper[4814]: I0227 16:59:24.030713 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:24 crc kubenswrapper[4814]: I0227 16:59:24.081630 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:24 crc kubenswrapper[4814]: I0227 16:59:24.288447 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r9q8w"] Feb 27 16:59:25 crc kubenswrapper[4814]: I0227 16:59:25.338424 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r9q8w" podUID="a0b4e818-ea6c-4450-be75-44c769aeb159" containerName="registry-server" containerID="cri-o://786b544f3b74635eda559cb5099a3b686bc38e1bfbd78c94a282086b49814f8b" gracePeriod=2 Feb 27 16:59:25 crc kubenswrapper[4814]: I0227 16:59:25.895402 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.026425 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0b4e818-ea6c-4450-be75-44c769aeb159-catalog-content\") pod \"a0b4e818-ea6c-4450-be75-44c769aeb159\" (UID: \"a0b4e818-ea6c-4450-be75-44c769aeb159\") " Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.026570 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fh96j\" (UniqueName: \"kubernetes.io/projected/a0b4e818-ea6c-4450-be75-44c769aeb159-kube-api-access-fh96j\") pod \"a0b4e818-ea6c-4450-be75-44c769aeb159\" (UID: \"a0b4e818-ea6c-4450-be75-44c769aeb159\") " Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.026636 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0b4e818-ea6c-4450-be75-44c769aeb159-utilities\") pod \"a0b4e818-ea6c-4450-be75-44c769aeb159\" (UID: \"a0b4e818-ea6c-4450-be75-44c769aeb159\") " Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.027911 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0b4e818-ea6c-4450-be75-44c769aeb159-utilities" (OuterVolumeSpecName: "utilities") pod "a0b4e818-ea6c-4450-be75-44c769aeb159" (UID: "a0b4e818-ea6c-4450-be75-44c769aeb159"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.035653 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0b4e818-ea6c-4450-be75-44c769aeb159-kube-api-access-fh96j" (OuterVolumeSpecName: "kube-api-access-fh96j") pod "a0b4e818-ea6c-4450-be75-44c769aeb159" (UID: "a0b4e818-ea6c-4450-be75-44c769aeb159"). InnerVolumeSpecName "kube-api-access-fh96j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.128810 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fh96j\" (UniqueName: \"kubernetes.io/projected/a0b4e818-ea6c-4450-be75-44c769aeb159-kube-api-access-fh96j\") on node \"crc\" DevicePath \"\"" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.128846 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0b4e818-ea6c-4450-be75-44c769aeb159-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.180772 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0b4e818-ea6c-4450-be75-44c769aeb159-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0b4e818-ea6c-4450-be75-44c769aeb159" (UID: "a0b4e818-ea6c-4450-be75-44c769aeb159"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.230195 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0b4e818-ea6c-4450-be75-44c769aeb159-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.358953 4814 generic.go:334] "Generic (PLEG): container finished" podID="a0b4e818-ea6c-4450-be75-44c769aeb159" containerID="786b544f3b74635eda559cb5099a3b686bc38e1bfbd78c94a282086b49814f8b" exitCode=0 Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.359045 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9q8w" event={"ID":"a0b4e818-ea6c-4450-be75-44c769aeb159","Type":"ContainerDied","Data":"786b544f3b74635eda559cb5099a3b686bc38e1bfbd78c94a282086b49814f8b"} Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.359089 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r9q8w" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.359124 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9q8w" event={"ID":"a0b4e818-ea6c-4450-be75-44c769aeb159","Type":"ContainerDied","Data":"9e8bc408074eaed3c9946d2e7531e0ba618b9b78ed6493ddf7e72f76e286d0ce"} Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.359165 4814 scope.go:117] "RemoveContainer" containerID="786b544f3b74635eda559cb5099a3b686bc38e1bfbd78c94a282086b49814f8b" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.409296 4814 scope.go:117] "RemoveContainer" containerID="43678db89c371aae86da7c8239fb05f4b3603821af79c149f414218c828d3f48" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.424116 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r9q8w"] Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.435708 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r9q8w"] Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.445808 4814 scope.go:117] "RemoveContainer" containerID="6a68c95c4aedca638d932bacacdc9074568a7bb79c0f52cd2b928cf3acf5dfac" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.487183 4814 scope.go:117] "RemoveContainer" containerID="786b544f3b74635eda559cb5099a3b686bc38e1bfbd78c94a282086b49814f8b" Feb 27 16:59:26 crc kubenswrapper[4814]: E0227 16:59:26.496882 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"786b544f3b74635eda559cb5099a3b686bc38e1bfbd78c94a282086b49814f8b\": container with ID starting with 786b544f3b74635eda559cb5099a3b686bc38e1bfbd78c94a282086b49814f8b not found: ID does not exist" containerID="786b544f3b74635eda559cb5099a3b686bc38e1bfbd78c94a282086b49814f8b" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.496936 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"786b544f3b74635eda559cb5099a3b686bc38e1bfbd78c94a282086b49814f8b"} err="failed to get container status \"786b544f3b74635eda559cb5099a3b686bc38e1bfbd78c94a282086b49814f8b\": rpc error: code = NotFound desc = could not find container \"786b544f3b74635eda559cb5099a3b686bc38e1bfbd78c94a282086b49814f8b\": container with ID starting with 786b544f3b74635eda559cb5099a3b686bc38e1bfbd78c94a282086b49814f8b not found: ID does not exist" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.496970 4814 scope.go:117] "RemoveContainer" containerID="43678db89c371aae86da7c8239fb05f4b3603821af79c149f414218c828d3f48" Feb 27 16:59:26 crc kubenswrapper[4814]: E0227 16:59:26.497401 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43678db89c371aae86da7c8239fb05f4b3603821af79c149f414218c828d3f48\": container with ID starting with 43678db89c371aae86da7c8239fb05f4b3603821af79c149f414218c828d3f48 not found: ID does not exist" containerID="43678db89c371aae86da7c8239fb05f4b3603821af79c149f414218c828d3f48" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.497467 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43678db89c371aae86da7c8239fb05f4b3603821af79c149f414218c828d3f48"} err="failed to get container status \"43678db89c371aae86da7c8239fb05f4b3603821af79c149f414218c828d3f48\": rpc error: code = NotFound desc = could not find container \"43678db89c371aae86da7c8239fb05f4b3603821af79c149f414218c828d3f48\": container with ID starting with 43678db89c371aae86da7c8239fb05f4b3603821af79c149f414218c828d3f48 not found: ID does not exist" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.497507 4814 scope.go:117] "RemoveContainer" containerID="6a68c95c4aedca638d932bacacdc9074568a7bb79c0f52cd2b928cf3acf5dfac" Feb 27 16:59:26 crc kubenswrapper[4814]: E0227 16:59:26.497914 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a68c95c4aedca638d932bacacdc9074568a7bb79c0f52cd2b928cf3acf5dfac\": container with ID starting with 6a68c95c4aedca638d932bacacdc9074568a7bb79c0f52cd2b928cf3acf5dfac not found: ID does not exist" containerID="6a68c95c4aedca638d932bacacdc9074568a7bb79c0f52cd2b928cf3acf5dfac" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.497963 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a68c95c4aedca638d932bacacdc9074568a7bb79c0f52cd2b928cf3acf5dfac"} err="failed to get container status \"6a68c95c4aedca638d932bacacdc9074568a7bb79c0f52cd2b928cf3acf5dfac\": rpc error: code = NotFound desc = could not find container \"6a68c95c4aedca638d932bacacdc9074568a7bb79c0f52cd2b928cf3acf5dfac\": container with ID starting with 6a68c95c4aedca638d932bacacdc9074568a7bb79c0f52cd2b928cf3acf5dfac not found: ID does not exist" Feb 27 16:59:26 crc kubenswrapper[4814]: I0227 16:59:26.506079 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0b4e818-ea6c-4450-be75-44c769aeb159" path="/var/lib/kubelet/pods/a0b4e818-ea6c-4450-be75-44c769aeb159/volumes" Feb 27 16:59:49 crc kubenswrapper[4814]: I0227 16:59:49.592487 4814 generic.go:334] "Generic (PLEG): container finished" podID="cf70174d-7319-418e-8545-83db61a7eb7c" containerID="18d10b2dc0d47e66434ca4914307de01a647006be6f3c1c0e2492fd3ce62d321" exitCode=0 Feb 27 16:59:49 crc kubenswrapper[4814]: I0227 16:59:49.592599 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" event={"ID":"cf70174d-7319-418e-8545-83db61a7eb7c","Type":"ContainerDied","Data":"18d10b2dc0d47e66434ca4914307de01a647006be6f3c1c0e2492fd3ce62d321"} Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.139995 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.244493 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"cf70174d-7319-418e-8545-83db61a7eb7c\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.244569 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-inventory\") pod \"cf70174d-7319-418e-8545-83db61a7eb7c\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.244727 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-neutron-metadata-combined-ca-bundle\") pod \"cf70174d-7319-418e-8545-83db61a7eb7c\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.244791 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-ssh-key-openstack-edpm-ipam\") pod \"cf70174d-7319-418e-8545-83db61a7eb7c\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.244843 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-nova-metadata-neutron-config-0\") pod \"cf70174d-7319-418e-8545-83db61a7eb7c\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.244892 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkgk6\" (UniqueName: \"kubernetes.io/projected/cf70174d-7319-418e-8545-83db61a7eb7c-kube-api-access-pkgk6\") pod \"cf70174d-7319-418e-8545-83db61a7eb7c\" (UID: \"cf70174d-7319-418e-8545-83db61a7eb7c\") " Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.253356 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf70174d-7319-418e-8545-83db61a7eb7c-kube-api-access-pkgk6" (OuterVolumeSpecName: "kube-api-access-pkgk6") pod "cf70174d-7319-418e-8545-83db61a7eb7c" (UID: "cf70174d-7319-418e-8545-83db61a7eb7c"). InnerVolumeSpecName "kube-api-access-pkgk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.253958 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "cf70174d-7319-418e-8545-83db61a7eb7c" (UID: "cf70174d-7319-418e-8545-83db61a7eb7c"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.288967 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "cf70174d-7319-418e-8545-83db61a7eb7c" (UID: "cf70174d-7319-418e-8545-83db61a7eb7c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.294931 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-inventory" (OuterVolumeSpecName: "inventory") pod "cf70174d-7319-418e-8545-83db61a7eb7c" (UID: "cf70174d-7319-418e-8545-83db61a7eb7c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.298360 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "cf70174d-7319-418e-8545-83db61a7eb7c" (UID: "cf70174d-7319-418e-8545-83db61a7eb7c"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.301617 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "cf70174d-7319-418e-8545-83db61a7eb7c" (UID: "cf70174d-7319-418e-8545-83db61a7eb7c"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.347331 4814 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.347375 4814 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.347390 4814 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.347406 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.347419 4814 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cf70174d-7319-418e-8545-83db61a7eb7c-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.347432 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkgk6\" (UniqueName: \"kubernetes.io/projected/cf70174d-7319-418e-8545-83db61a7eb7c-kube-api-access-pkgk6\") on node \"crc\" DevicePath \"\"" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.636195 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" event={"ID":"cf70174d-7319-418e-8545-83db61a7eb7c","Type":"ContainerDied","Data":"66d594822bde897c5cb075c5611d94898b13ece518e899d539c9852f5d5101a7"} Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.636284 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.636285 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66d594822bde897c5cb075c5611d94898b13ece518e899d539c9852f5d5101a7" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.839574 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n"] Feb 27 16:59:51 crc kubenswrapper[4814]: E0227 16:59:51.840000 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b4e818-ea6c-4450-be75-44c769aeb159" containerName="extract-utilities" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.840027 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b4e818-ea6c-4450-be75-44c769aeb159" containerName="extract-utilities" Feb 27 16:59:51 crc kubenswrapper[4814]: E0227 16:59:51.840047 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b4e818-ea6c-4450-be75-44c769aeb159" containerName="registry-server" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.840058 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b4e818-ea6c-4450-be75-44c769aeb159" containerName="registry-server" Feb 27 16:59:51 crc kubenswrapper[4814]: E0227 16:59:51.840096 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf70174d-7319-418e-8545-83db61a7eb7c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.840111 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf70174d-7319-418e-8545-83db61a7eb7c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 27 16:59:51 crc kubenswrapper[4814]: E0227 16:59:51.840133 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b4e818-ea6c-4450-be75-44c769aeb159" containerName="extract-content" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.840144 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b4e818-ea6c-4450-be75-44c769aeb159" containerName="extract-content" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.840458 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf70174d-7319-418e-8545-83db61a7eb7c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.840484 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b4e818-ea6c-4450-be75-44c769aeb159" containerName="registry-server" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.841191 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.846468 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.846553 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.846584 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.848595 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.848978 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.856739 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n"] Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.967622 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.967674 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.967712 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.967776 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:51 crc kubenswrapper[4814]: I0227 16:59:51.967884 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn7vx\" (UniqueName: \"kubernetes.io/projected/c96b1d1f-6478-4eea-8413-a0657f19c6e7-kube-api-access-zn7vx\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:52 crc kubenswrapper[4814]: I0227 16:59:52.069415 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn7vx\" (UniqueName: \"kubernetes.io/projected/c96b1d1f-6478-4eea-8413-a0657f19c6e7-kube-api-access-zn7vx\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:52 crc kubenswrapper[4814]: I0227 16:59:52.069547 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:52 crc kubenswrapper[4814]: I0227 16:59:52.069597 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:52 crc kubenswrapper[4814]: I0227 16:59:52.070376 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:52 crc kubenswrapper[4814]: I0227 16:59:52.070511 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:52 crc kubenswrapper[4814]: I0227 16:59:52.075891 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:52 crc kubenswrapper[4814]: I0227 16:59:52.076315 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:52 crc kubenswrapper[4814]: I0227 16:59:52.076470 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:52 crc kubenswrapper[4814]: I0227 16:59:52.079244 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:52 crc kubenswrapper[4814]: I0227 16:59:52.086540 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn7vx\" (UniqueName: \"kubernetes.io/projected/c96b1d1f-6478-4eea-8413-a0657f19c6e7-kube-api-access-zn7vx\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:52 crc kubenswrapper[4814]: I0227 16:59:52.158303 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 16:59:52 crc kubenswrapper[4814]: I0227 16:59:52.762466 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n"] Feb 27 16:59:53 crc kubenswrapper[4814]: I0227 16:59:53.664572 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" event={"ID":"c96b1d1f-6478-4eea-8413-a0657f19c6e7","Type":"ContainerStarted","Data":"35730396da47ff0b822a5f1a5840f41d14871ea95fb1f4cc78ff307f88a3d6ed"} Feb 27 16:59:53 crc kubenswrapper[4814]: I0227 16:59:53.665354 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" event={"ID":"c96b1d1f-6478-4eea-8413-a0657f19c6e7","Type":"ContainerStarted","Data":"b7dc90de91b166e4b7980aa4047accbea01899b440981816ef3760455e011130"} Feb 27 16:59:53 crc kubenswrapper[4814]: I0227 16:59:53.703740 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" podStartSLOduration=2.29464509 podStartE2EDuration="2.703714269s" podCreationTimestamp="2026-02-27 16:59:51 +0000 UTC" firstStartedPulling="2026-02-27 16:59:52.778646193 +0000 UTC m=+2205.231271083" lastFinishedPulling="2026-02-27 16:59:53.187715402 +0000 UTC m=+2205.640340262" observedRunningTime="2026-02-27 16:59:53.693398821 +0000 UTC m=+2206.146023691" watchObservedRunningTime="2026-02-27 16:59:53.703714269 +0000 UTC m=+2206.156339129" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.143314 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536860-n5hhd"] Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.145895 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536860-n5hhd" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.149939 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.150595 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.152718 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.159779 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536860-n5hhd"] Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.171972 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5"] Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.173957 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.176361 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.177928 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.183832 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5"] Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.273173 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk54t\" (UniqueName: \"kubernetes.io/projected/d7ddb784-229e-408f-988b-c68f84154eb7-kube-api-access-dk54t\") pod \"collect-profiles-29536860-scsg5\" (UID: \"d7ddb784-229e-408f-988b-c68f84154eb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.273305 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d7ddb784-229e-408f-988b-c68f84154eb7-secret-volume\") pod \"collect-profiles-29536860-scsg5\" (UID: \"d7ddb784-229e-408f-988b-c68f84154eb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.273670 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d7ddb784-229e-408f-988b-c68f84154eb7-config-volume\") pod \"collect-profiles-29536860-scsg5\" (UID: \"d7ddb784-229e-408f-988b-c68f84154eb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.273829 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55swk\" (UniqueName: \"kubernetes.io/projected/8879d490-f906-4e15-980b-a4bedc4a3196-kube-api-access-55swk\") pod \"auto-csr-approver-29536860-n5hhd\" (UID: \"8879d490-f906-4e15-980b-a4bedc4a3196\") " pod="openshift-infra/auto-csr-approver-29536860-n5hhd" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.376432 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d7ddb784-229e-408f-988b-c68f84154eb7-secret-volume\") pod \"collect-profiles-29536860-scsg5\" (UID: \"d7ddb784-229e-408f-988b-c68f84154eb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.376641 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d7ddb784-229e-408f-988b-c68f84154eb7-config-volume\") pod \"collect-profiles-29536860-scsg5\" (UID: \"d7ddb784-229e-408f-988b-c68f84154eb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.376730 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55swk\" (UniqueName: \"kubernetes.io/projected/8879d490-f906-4e15-980b-a4bedc4a3196-kube-api-access-55swk\") pod \"auto-csr-approver-29536860-n5hhd\" (UID: \"8879d490-f906-4e15-980b-a4bedc4a3196\") " pod="openshift-infra/auto-csr-approver-29536860-n5hhd" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.376828 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk54t\" (UniqueName: \"kubernetes.io/projected/d7ddb784-229e-408f-988b-c68f84154eb7-kube-api-access-dk54t\") pod \"collect-profiles-29536860-scsg5\" (UID: \"d7ddb784-229e-408f-988b-c68f84154eb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.378870 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d7ddb784-229e-408f-988b-c68f84154eb7-config-volume\") pod \"collect-profiles-29536860-scsg5\" (UID: \"d7ddb784-229e-408f-988b-c68f84154eb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.385734 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d7ddb784-229e-408f-988b-c68f84154eb7-secret-volume\") pod \"collect-profiles-29536860-scsg5\" (UID: \"d7ddb784-229e-408f-988b-c68f84154eb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.417671 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk54t\" (UniqueName: \"kubernetes.io/projected/d7ddb784-229e-408f-988b-c68f84154eb7-kube-api-access-dk54t\") pod \"collect-profiles-29536860-scsg5\" (UID: \"d7ddb784-229e-408f-988b-c68f84154eb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.418115 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55swk\" (UniqueName: \"kubernetes.io/projected/8879d490-f906-4e15-980b-a4bedc4a3196-kube-api-access-55swk\") pod \"auto-csr-approver-29536860-n5hhd\" (UID: \"8879d490-f906-4e15-980b-a4bedc4a3196\") " pod="openshift-infra/auto-csr-approver-29536860-n5hhd" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.513952 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536860-n5hhd" Feb 27 17:00:00 crc kubenswrapper[4814]: I0227 17:00:00.540984 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5" Feb 27 17:00:01 crc kubenswrapper[4814]: I0227 17:00:01.014960 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536860-n5hhd"] Feb 27 17:00:01 crc kubenswrapper[4814]: W0227 17:00:01.023741 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8879d490_f906_4e15_980b_a4bedc4a3196.slice/crio-4e4fbb579b51fdfb26f53d87cd5b47374bad97dbe2dd04ed83a8a0723ecd8a61 WatchSource:0}: Error finding container 4e4fbb579b51fdfb26f53d87cd5b47374bad97dbe2dd04ed83a8a0723ecd8a61: Status 404 returned error can't find the container with id 4e4fbb579b51fdfb26f53d87cd5b47374bad97dbe2dd04ed83a8a0723ecd8a61 Feb 27 17:00:01 crc kubenswrapper[4814]: W0227 17:00:01.109123 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7ddb784_229e_408f_988b_c68f84154eb7.slice/crio-40e0ded82e43c6b1062b62f72c96ec0d97c7ca0bcd6b9897fe02377e9f4157f1 WatchSource:0}: Error finding container 40e0ded82e43c6b1062b62f72c96ec0d97c7ca0bcd6b9897fe02377e9f4157f1: Status 404 returned error can't find the container with id 40e0ded82e43c6b1062b62f72c96ec0d97c7ca0bcd6b9897fe02377e9f4157f1 Feb 27 17:00:01 crc kubenswrapper[4814]: I0227 17:00:01.117325 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5"] Feb 27 17:00:01 crc kubenswrapper[4814]: I0227 17:00:01.753227 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536860-n5hhd" event={"ID":"8879d490-f906-4e15-980b-a4bedc4a3196","Type":"ContainerStarted","Data":"4e4fbb579b51fdfb26f53d87cd5b47374bad97dbe2dd04ed83a8a0723ecd8a61"} Feb 27 17:00:01 crc kubenswrapper[4814]: I0227 17:00:01.756911 4814 generic.go:334] "Generic (PLEG): container finished" podID="d7ddb784-229e-408f-988b-c68f84154eb7" containerID="16b6ba44f291ab7470bfa3810314209e8c144cb8ae32773b088f6f1387a14b17" exitCode=0 Feb 27 17:00:01 crc kubenswrapper[4814]: I0227 17:00:01.756974 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5" event={"ID":"d7ddb784-229e-408f-988b-c68f84154eb7","Type":"ContainerDied","Data":"16b6ba44f291ab7470bfa3810314209e8c144cb8ae32773b088f6f1387a14b17"} Feb 27 17:00:01 crc kubenswrapper[4814]: I0227 17:00:01.757347 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5" event={"ID":"d7ddb784-229e-408f-988b-c68f84154eb7","Type":"ContainerStarted","Data":"40e0ded82e43c6b1062b62f72c96ec0d97c7ca0bcd6b9897fe02377e9f4157f1"} Feb 27 17:00:03 crc kubenswrapper[4814]: I0227 17:00:03.235829 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5" Feb 27 17:00:03 crc kubenswrapper[4814]: I0227 17:00:03.339681 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk54t\" (UniqueName: \"kubernetes.io/projected/d7ddb784-229e-408f-988b-c68f84154eb7-kube-api-access-dk54t\") pod \"d7ddb784-229e-408f-988b-c68f84154eb7\" (UID: \"d7ddb784-229e-408f-988b-c68f84154eb7\") " Feb 27 17:00:03 crc kubenswrapper[4814]: I0227 17:00:03.339820 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d7ddb784-229e-408f-988b-c68f84154eb7-config-volume\") pod \"d7ddb784-229e-408f-988b-c68f84154eb7\" (UID: \"d7ddb784-229e-408f-988b-c68f84154eb7\") " Feb 27 17:00:03 crc kubenswrapper[4814]: I0227 17:00:03.339969 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d7ddb784-229e-408f-988b-c68f84154eb7-secret-volume\") pod \"d7ddb784-229e-408f-988b-c68f84154eb7\" (UID: \"d7ddb784-229e-408f-988b-c68f84154eb7\") " Feb 27 17:00:03 crc kubenswrapper[4814]: I0227 17:00:03.341207 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7ddb784-229e-408f-988b-c68f84154eb7-config-volume" (OuterVolumeSpecName: "config-volume") pod "d7ddb784-229e-408f-988b-c68f84154eb7" (UID: "d7ddb784-229e-408f-988b-c68f84154eb7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:00:03 crc kubenswrapper[4814]: I0227 17:00:03.346494 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7ddb784-229e-408f-988b-c68f84154eb7-kube-api-access-dk54t" (OuterVolumeSpecName: "kube-api-access-dk54t") pod "d7ddb784-229e-408f-988b-c68f84154eb7" (UID: "d7ddb784-229e-408f-988b-c68f84154eb7"). InnerVolumeSpecName "kube-api-access-dk54t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:00:03 crc kubenswrapper[4814]: I0227 17:00:03.347160 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7ddb784-229e-408f-988b-c68f84154eb7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d7ddb784-229e-408f-988b-c68f84154eb7" (UID: "d7ddb784-229e-408f-988b-c68f84154eb7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:00:03 crc kubenswrapper[4814]: I0227 17:00:03.442858 4814 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d7ddb784-229e-408f-988b-c68f84154eb7-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 17:00:03 crc kubenswrapper[4814]: I0227 17:00:03.442898 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk54t\" (UniqueName: \"kubernetes.io/projected/d7ddb784-229e-408f-988b-c68f84154eb7-kube-api-access-dk54t\") on node \"crc\" DevicePath \"\"" Feb 27 17:00:03 crc kubenswrapper[4814]: I0227 17:00:03.442908 4814 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d7ddb784-229e-408f-988b-c68f84154eb7-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 17:00:03 crc kubenswrapper[4814]: I0227 17:00:03.784045 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5" event={"ID":"d7ddb784-229e-408f-988b-c68f84154eb7","Type":"ContainerDied","Data":"40e0ded82e43c6b1062b62f72c96ec0d97c7ca0bcd6b9897fe02377e9f4157f1"} Feb 27 17:00:03 crc kubenswrapper[4814]: I0227 17:00:03.784438 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40e0ded82e43c6b1062b62f72c96ec0d97c7ca0bcd6b9897fe02377e9f4157f1" Feb 27 17:00:03 crc kubenswrapper[4814]: I0227 17:00:03.784127 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-scsg5" Feb 27 17:00:04 crc kubenswrapper[4814]: I0227 17:00:04.350513 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb"] Feb 27 17:00:04 crc kubenswrapper[4814]: I0227 17:00:04.360979 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536815-pqwsb"] Feb 27 17:00:04 crc kubenswrapper[4814]: I0227 17:00:04.506458 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d87f2169-8a70-42de-aaf5-7728ff95fa50" path="/var/lib/kubelet/pods/d87f2169-8a70-42de-aaf5-7728ff95fa50/volumes" Feb 27 17:00:07 crc kubenswrapper[4814]: I0227 17:00:07.843026 4814 generic.go:334] "Generic (PLEG): container finished" podID="8879d490-f906-4e15-980b-a4bedc4a3196" containerID="a1980c7a0753fa4ef15800e06dd6c6ed370cf7b727c402ba2224547849faac7a" exitCode=0 Feb 27 17:00:07 crc kubenswrapper[4814]: I0227 17:00:07.843463 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536860-n5hhd" event={"ID":"8879d490-f906-4e15-980b-a4bedc4a3196","Type":"ContainerDied","Data":"a1980c7a0753fa4ef15800e06dd6c6ed370cf7b727c402ba2224547849faac7a"} Feb 27 17:00:09 crc kubenswrapper[4814]: I0227 17:00:09.236149 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536860-n5hhd" Feb 27 17:00:09 crc kubenswrapper[4814]: I0227 17:00:09.398588 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55swk\" (UniqueName: \"kubernetes.io/projected/8879d490-f906-4e15-980b-a4bedc4a3196-kube-api-access-55swk\") pod \"8879d490-f906-4e15-980b-a4bedc4a3196\" (UID: \"8879d490-f906-4e15-980b-a4bedc4a3196\") " Feb 27 17:00:09 crc kubenswrapper[4814]: I0227 17:00:09.404312 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8879d490-f906-4e15-980b-a4bedc4a3196-kube-api-access-55swk" (OuterVolumeSpecName: "kube-api-access-55swk") pod "8879d490-f906-4e15-980b-a4bedc4a3196" (UID: "8879d490-f906-4e15-980b-a4bedc4a3196"). InnerVolumeSpecName "kube-api-access-55swk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:00:09 crc kubenswrapper[4814]: I0227 17:00:09.501189 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55swk\" (UniqueName: \"kubernetes.io/projected/8879d490-f906-4e15-980b-a4bedc4a3196-kube-api-access-55swk\") on node \"crc\" DevicePath \"\"" Feb 27 17:00:09 crc kubenswrapper[4814]: I0227 17:00:09.863027 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536860-n5hhd" event={"ID":"8879d490-f906-4e15-980b-a4bedc4a3196","Type":"ContainerDied","Data":"4e4fbb579b51fdfb26f53d87cd5b47374bad97dbe2dd04ed83a8a0723ecd8a61"} Feb 27 17:00:09 crc kubenswrapper[4814]: I0227 17:00:09.863084 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e4fbb579b51fdfb26f53d87cd5b47374bad97dbe2dd04ed83a8a0723ecd8a61" Feb 27 17:00:09 crc kubenswrapper[4814]: I0227 17:00:09.863173 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536860-n5hhd" Feb 27 17:00:10 crc kubenswrapper[4814]: I0227 17:00:10.307713 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536854-pkbjj"] Feb 27 17:00:10 crc kubenswrapper[4814]: I0227 17:00:10.318512 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536854-pkbjj"] Feb 27 17:00:10 crc kubenswrapper[4814]: I0227 17:00:10.509215 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d93ea82-2104-4731-93ae-3e469a251396" path="/var/lib/kubelet/pods/3d93ea82-2104-4731-93ae-3e469a251396/volumes" Feb 27 17:00:22 crc kubenswrapper[4814]: I0227 17:00:22.903309 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:00:22 crc kubenswrapper[4814]: I0227 17:00:22.906218 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:00:52 crc kubenswrapper[4814]: I0227 17:00:52.903128 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:00:52 crc kubenswrapper[4814]: I0227 17:00:52.903978 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:00:56 crc kubenswrapper[4814]: I0227 17:00:56.820066 4814 scope.go:117] "RemoveContainer" containerID="1b66aa8513771120b04a51a53cb2e88cc6798d2624ffbbdcc62fc18f82ad397e" Feb 27 17:00:56 crc kubenswrapper[4814]: I0227 17:00:56.890250 4814 scope.go:117] "RemoveContainer" containerID="b86ccc9ec3b77950f4a33f3517ef3218c0cf88bf4a5d8200fb3cda18db954532" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.171205 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29536861-pnnbc"] Feb 27 17:01:00 crc kubenswrapper[4814]: E0227 17:01:00.172680 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ddb784-229e-408f-988b-c68f84154eb7" containerName="collect-profiles" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.172715 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ddb784-229e-408f-988b-c68f84154eb7" containerName="collect-profiles" Feb 27 17:01:00 crc kubenswrapper[4814]: E0227 17:01:00.172777 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8879d490-f906-4e15-980b-a4bedc4a3196" containerName="oc" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.172793 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="8879d490-f906-4e15-980b-a4bedc4a3196" containerName="oc" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.173227 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="8879d490-f906-4e15-980b-a4bedc4a3196" containerName="oc" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.173331 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7ddb784-229e-408f-988b-c68f84154eb7" containerName="collect-profiles" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.174876 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29536861-pnnbc" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.194017 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29536861-pnnbc"] Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.216707 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-combined-ca-bundle\") pod \"keystone-cron-29536861-pnnbc\" (UID: \"24ebfded-b93e-4720-9747-75b8947d26c1\") " pod="openstack/keystone-cron-29536861-pnnbc" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.216766 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-fernet-keys\") pod \"keystone-cron-29536861-pnnbc\" (UID: \"24ebfded-b93e-4720-9747-75b8947d26c1\") " pod="openstack/keystone-cron-29536861-pnnbc" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.216830 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-config-data\") pod \"keystone-cron-29536861-pnnbc\" (UID: \"24ebfded-b93e-4720-9747-75b8947d26c1\") " pod="openstack/keystone-cron-29536861-pnnbc" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.216906 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7xk5\" (UniqueName: \"kubernetes.io/projected/24ebfded-b93e-4720-9747-75b8947d26c1-kube-api-access-z7xk5\") pod \"keystone-cron-29536861-pnnbc\" (UID: \"24ebfded-b93e-4720-9747-75b8947d26c1\") " pod="openstack/keystone-cron-29536861-pnnbc" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.318818 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7xk5\" (UniqueName: \"kubernetes.io/projected/24ebfded-b93e-4720-9747-75b8947d26c1-kube-api-access-z7xk5\") pod \"keystone-cron-29536861-pnnbc\" (UID: \"24ebfded-b93e-4720-9747-75b8947d26c1\") " pod="openstack/keystone-cron-29536861-pnnbc" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.319053 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-combined-ca-bundle\") pod \"keystone-cron-29536861-pnnbc\" (UID: \"24ebfded-b93e-4720-9747-75b8947d26c1\") " pod="openstack/keystone-cron-29536861-pnnbc" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.319104 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-fernet-keys\") pod \"keystone-cron-29536861-pnnbc\" (UID: \"24ebfded-b93e-4720-9747-75b8947d26c1\") " pod="openstack/keystone-cron-29536861-pnnbc" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.319249 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-config-data\") pod \"keystone-cron-29536861-pnnbc\" (UID: \"24ebfded-b93e-4720-9747-75b8947d26c1\") " pod="openstack/keystone-cron-29536861-pnnbc" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.334306 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-fernet-keys\") pod \"keystone-cron-29536861-pnnbc\" (UID: \"24ebfded-b93e-4720-9747-75b8947d26c1\") " pod="openstack/keystone-cron-29536861-pnnbc" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.339033 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-config-data\") pod \"keystone-cron-29536861-pnnbc\" (UID: \"24ebfded-b93e-4720-9747-75b8947d26c1\") " pod="openstack/keystone-cron-29536861-pnnbc" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.340333 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-combined-ca-bundle\") pod \"keystone-cron-29536861-pnnbc\" (UID: \"24ebfded-b93e-4720-9747-75b8947d26c1\") " pod="openstack/keystone-cron-29536861-pnnbc" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.347590 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7xk5\" (UniqueName: \"kubernetes.io/projected/24ebfded-b93e-4720-9747-75b8947d26c1-kube-api-access-z7xk5\") pod \"keystone-cron-29536861-pnnbc\" (UID: \"24ebfded-b93e-4720-9747-75b8947d26c1\") " pod="openstack/keystone-cron-29536861-pnnbc" Feb 27 17:01:00 crc kubenswrapper[4814]: I0227 17:01:00.498878 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29536861-pnnbc" Feb 27 17:01:01 crc kubenswrapper[4814]: I0227 17:01:01.001424 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29536861-pnnbc"] Feb 27 17:01:01 crc kubenswrapper[4814]: I0227 17:01:01.415987 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29536861-pnnbc" event={"ID":"24ebfded-b93e-4720-9747-75b8947d26c1","Type":"ContainerStarted","Data":"ffad672596dd46f1029a157a9e201328bf747a7c04df9d5196639977bcca4bd7"} Feb 27 17:01:01 crc kubenswrapper[4814]: I0227 17:01:01.416038 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29536861-pnnbc" event={"ID":"24ebfded-b93e-4720-9747-75b8947d26c1","Type":"ContainerStarted","Data":"b058cf979a528c96667f036b1ebaa16945d2ab4869b3404af929bf422dcb6118"} Feb 27 17:01:01 crc kubenswrapper[4814]: I0227 17:01:01.438069 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29536861-pnnbc" podStartSLOduration=1.438054156 podStartE2EDuration="1.438054156s" podCreationTimestamp="2026-02-27 17:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:01:01.434993192 +0000 UTC m=+2273.887618012" watchObservedRunningTime="2026-02-27 17:01:01.438054156 +0000 UTC m=+2273.890678986" Feb 27 17:01:03 crc kubenswrapper[4814]: I0227 17:01:03.442920 4814 generic.go:334] "Generic (PLEG): container finished" podID="24ebfded-b93e-4720-9747-75b8947d26c1" containerID="ffad672596dd46f1029a157a9e201328bf747a7c04df9d5196639977bcca4bd7" exitCode=0 Feb 27 17:01:03 crc kubenswrapper[4814]: I0227 17:01:03.443010 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29536861-pnnbc" event={"ID":"24ebfded-b93e-4720-9747-75b8947d26c1","Type":"ContainerDied","Data":"ffad672596dd46f1029a157a9e201328bf747a7c04df9d5196639977bcca4bd7"} Feb 27 17:01:04 crc kubenswrapper[4814]: I0227 17:01:04.116769 4814 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-69895c5d4f-4bpj5" podUID="9e55471d-9b78-4e06-8195-c73f4ce0d1f6" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Feb 27 17:01:04 crc kubenswrapper[4814]: I0227 17:01:04.833834 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29536861-pnnbc" Feb 27 17:01:04 crc kubenswrapper[4814]: I0227 17:01:04.927274 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-combined-ca-bundle\") pod \"24ebfded-b93e-4720-9747-75b8947d26c1\" (UID: \"24ebfded-b93e-4720-9747-75b8947d26c1\") " Feb 27 17:01:04 crc kubenswrapper[4814]: I0227 17:01:04.984418 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24ebfded-b93e-4720-9747-75b8947d26c1" (UID: "24ebfded-b93e-4720-9747-75b8947d26c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:01:05 crc kubenswrapper[4814]: I0227 17:01:05.028826 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7xk5\" (UniqueName: \"kubernetes.io/projected/24ebfded-b93e-4720-9747-75b8947d26c1-kube-api-access-z7xk5\") pod \"24ebfded-b93e-4720-9747-75b8947d26c1\" (UID: \"24ebfded-b93e-4720-9747-75b8947d26c1\") " Feb 27 17:01:05 crc kubenswrapper[4814]: I0227 17:01:05.028870 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-config-data\") pod \"24ebfded-b93e-4720-9747-75b8947d26c1\" (UID: \"24ebfded-b93e-4720-9747-75b8947d26c1\") " Feb 27 17:01:05 crc kubenswrapper[4814]: I0227 17:01:05.028986 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-fernet-keys\") pod \"24ebfded-b93e-4720-9747-75b8947d26c1\" (UID: \"24ebfded-b93e-4720-9747-75b8947d26c1\") " Feb 27 17:01:05 crc kubenswrapper[4814]: I0227 17:01:05.029461 4814 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:01:05 crc kubenswrapper[4814]: I0227 17:01:05.033397 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "24ebfded-b93e-4720-9747-75b8947d26c1" (UID: "24ebfded-b93e-4720-9747-75b8947d26c1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:01:05 crc kubenswrapper[4814]: I0227 17:01:05.034304 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24ebfded-b93e-4720-9747-75b8947d26c1-kube-api-access-z7xk5" (OuterVolumeSpecName: "kube-api-access-z7xk5") pod "24ebfded-b93e-4720-9747-75b8947d26c1" (UID: "24ebfded-b93e-4720-9747-75b8947d26c1"). InnerVolumeSpecName "kube-api-access-z7xk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:01:05 crc kubenswrapper[4814]: I0227 17:01:05.081285 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-config-data" (OuterVolumeSpecName: "config-data") pod "24ebfded-b93e-4720-9747-75b8947d26c1" (UID: "24ebfded-b93e-4720-9747-75b8947d26c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:01:05 crc kubenswrapper[4814]: I0227 17:01:05.132150 4814 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 27 17:01:05 crc kubenswrapper[4814]: I0227 17:01:05.132198 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7xk5\" (UniqueName: \"kubernetes.io/projected/24ebfded-b93e-4720-9747-75b8947d26c1-kube-api-access-z7xk5\") on node \"crc\" DevicePath \"\"" Feb 27 17:01:05 crc kubenswrapper[4814]: I0227 17:01:05.132220 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ebfded-b93e-4720-9747-75b8947d26c1-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:01:05 crc kubenswrapper[4814]: I0227 17:01:05.467204 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29536861-pnnbc" event={"ID":"24ebfded-b93e-4720-9747-75b8947d26c1","Type":"ContainerDied","Data":"b058cf979a528c96667f036b1ebaa16945d2ab4869b3404af929bf422dcb6118"} Feb 27 17:01:05 crc kubenswrapper[4814]: I0227 17:01:05.467298 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b058cf979a528c96667f036b1ebaa16945d2ab4869b3404af929bf422dcb6118" Feb 27 17:01:05 crc kubenswrapper[4814]: I0227 17:01:05.467406 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29536861-pnnbc" Feb 27 17:01:22 crc kubenswrapper[4814]: I0227 17:01:22.902202 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:01:22 crc kubenswrapper[4814]: I0227 17:01:22.903079 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:01:22 crc kubenswrapper[4814]: I0227 17:01:22.903152 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 17:01:22 crc kubenswrapper[4814]: I0227 17:01:22.904616 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eb9a72219e9b318e493f16a1ab078d7373596a6f5f4b8c4c5a593b8ea470dc7c"} pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:01:22 crc kubenswrapper[4814]: I0227 17:01:22.904729 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" containerID="cri-o://eb9a72219e9b318e493f16a1ab078d7373596a6f5f4b8c4c5a593b8ea470dc7c" gracePeriod=600 Feb 27 17:01:23 crc kubenswrapper[4814]: I0227 17:01:23.994568 4814 generic.go:334] "Generic (PLEG): container finished" podID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerID="eb9a72219e9b318e493f16a1ab078d7373596a6f5f4b8c4c5a593b8ea470dc7c" exitCode=0 Feb 27 17:01:23 crc kubenswrapper[4814]: I0227 17:01:23.994641 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerDied","Data":"eb9a72219e9b318e493f16a1ab078d7373596a6f5f4b8c4c5a593b8ea470dc7c"} Feb 27 17:01:23 crc kubenswrapper[4814]: I0227 17:01:23.995390 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a"} Feb 27 17:01:23 crc kubenswrapper[4814]: I0227 17:01:23.995427 4814 scope.go:117] "RemoveContainer" containerID="aa50608cee78bbedba73d1882ddf6c28ab492701f6dddd22da0d6e445825da08" Feb 27 17:01:38 crc kubenswrapper[4814]: I0227 17:01:38.932404 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d8bvb"] Feb 27 17:01:38 crc kubenswrapper[4814]: E0227 17:01:38.933636 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24ebfded-b93e-4720-9747-75b8947d26c1" containerName="keystone-cron" Feb 27 17:01:38 crc kubenswrapper[4814]: I0227 17:01:38.933652 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="24ebfded-b93e-4720-9747-75b8947d26c1" containerName="keystone-cron" Feb 27 17:01:38 crc kubenswrapper[4814]: I0227 17:01:38.933893 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="24ebfded-b93e-4720-9747-75b8947d26c1" containerName="keystone-cron" Feb 27 17:01:38 crc kubenswrapper[4814]: I0227 17:01:38.935794 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:38 crc kubenswrapper[4814]: I0227 17:01:38.958556 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d8bvb"] Feb 27 17:01:39 crc kubenswrapper[4814]: I0227 17:01:39.109198 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c5cf1bf-c514-4f94-9663-0760495dbe57-utilities\") pod \"certified-operators-d8bvb\" (UID: \"8c5cf1bf-c514-4f94-9663-0760495dbe57\") " pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:39 crc kubenswrapper[4814]: I0227 17:01:39.109304 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72tsq\" (UniqueName: \"kubernetes.io/projected/8c5cf1bf-c514-4f94-9663-0760495dbe57-kube-api-access-72tsq\") pod \"certified-operators-d8bvb\" (UID: \"8c5cf1bf-c514-4f94-9663-0760495dbe57\") " pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:39 crc kubenswrapper[4814]: I0227 17:01:39.109394 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c5cf1bf-c514-4f94-9663-0760495dbe57-catalog-content\") pod \"certified-operators-d8bvb\" (UID: \"8c5cf1bf-c514-4f94-9663-0760495dbe57\") " pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:39 crc kubenswrapper[4814]: I0227 17:01:39.211649 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c5cf1bf-c514-4f94-9663-0760495dbe57-utilities\") pod \"certified-operators-d8bvb\" (UID: \"8c5cf1bf-c514-4f94-9663-0760495dbe57\") " pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:39 crc kubenswrapper[4814]: I0227 17:01:39.211739 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72tsq\" (UniqueName: \"kubernetes.io/projected/8c5cf1bf-c514-4f94-9663-0760495dbe57-kube-api-access-72tsq\") pod \"certified-operators-d8bvb\" (UID: \"8c5cf1bf-c514-4f94-9663-0760495dbe57\") " pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:39 crc kubenswrapper[4814]: I0227 17:01:39.211830 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c5cf1bf-c514-4f94-9663-0760495dbe57-catalog-content\") pod \"certified-operators-d8bvb\" (UID: \"8c5cf1bf-c514-4f94-9663-0760495dbe57\") " pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:39 crc kubenswrapper[4814]: I0227 17:01:39.212687 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c5cf1bf-c514-4f94-9663-0760495dbe57-utilities\") pod \"certified-operators-d8bvb\" (UID: \"8c5cf1bf-c514-4f94-9663-0760495dbe57\") " pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:39 crc kubenswrapper[4814]: I0227 17:01:39.212734 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c5cf1bf-c514-4f94-9663-0760495dbe57-catalog-content\") pod \"certified-operators-d8bvb\" (UID: \"8c5cf1bf-c514-4f94-9663-0760495dbe57\") " pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:39 crc kubenswrapper[4814]: I0227 17:01:39.258557 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72tsq\" (UniqueName: \"kubernetes.io/projected/8c5cf1bf-c514-4f94-9663-0760495dbe57-kube-api-access-72tsq\") pod \"certified-operators-d8bvb\" (UID: \"8c5cf1bf-c514-4f94-9663-0760495dbe57\") " pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:39 crc kubenswrapper[4814]: I0227 17:01:39.261096 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:39 crc kubenswrapper[4814]: I0227 17:01:39.731686 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d8bvb"] Feb 27 17:01:39 crc kubenswrapper[4814]: W0227 17:01:39.737015 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c5cf1bf_c514_4f94_9663_0760495dbe57.slice/crio-ed74c73f9dfc2d1594f695d7a09b9b861ef4f31a3ae6df280103ac46caf544d7 WatchSource:0}: Error finding container ed74c73f9dfc2d1594f695d7a09b9b861ef4f31a3ae6df280103ac46caf544d7: Status 404 returned error can't find the container with id ed74c73f9dfc2d1594f695d7a09b9b861ef4f31a3ae6df280103ac46caf544d7 Feb 27 17:01:40 crc kubenswrapper[4814]: I0227 17:01:40.184716 4814 generic.go:334] "Generic (PLEG): container finished" podID="8c5cf1bf-c514-4f94-9663-0760495dbe57" containerID="44fdcb5754c8fbf6f34ddc1c97287daf5786f9a2c5763f1e50954837724846eb" exitCode=0 Feb 27 17:01:40 crc kubenswrapper[4814]: I0227 17:01:40.184806 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8bvb" event={"ID":"8c5cf1bf-c514-4f94-9663-0760495dbe57","Type":"ContainerDied","Data":"44fdcb5754c8fbf6f34ddc1c97287daf5786f9a2c5763f1e50954837724846eb"} Feb 27 17:01:40 crc kubenswrapper[4814]: I0227 17:01:40.185238 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8bvb" event={"ID":"8c5cf1bf-c514-4f94-9663-0760495dbe57","Type":"ContainerStarted","Data":"ed74c73f9dfc2d1594f695d7a09b9b861ef4f31a3ae6df280103ac46caf544d7"} Feb 27 17:01:42 crc kubenswrapper[4814]: I0227 17:01:42.215893 4814 generic.go:334] "Generic (PLEG): container finished" podID="8c5cf1bf-c514-4f94-9663-0760495dbe57" containerID="458162105fbd39017477fe029904eb22e30723ea07aa9050b427a049342ca3c0" exitCode=0 Feb 27 17:01:42 crc kubenswrapper[4814]: I0227 17:01:42.215977 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8bvb" event={"ID":"8c5cf1bf-c514-4f94-9663-0760495dbe57","Type":"ContainerDied","Data":"458162105fbd39017477fe029904eb22e30723ea07aa9050b427a049342ca3c0"} Feb 27 17:01:43 crc kubenswrapper[4814]: I0227 17:01:43.246553 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8bvb" event={"ID":"8c5cf1bf-c514-4f94-9663-0760495dbe57","Type":"ContainerStarted","Data":"9cca9c0c274c438a18a0efe306f58b3e474ca0233accc1f0178d7f2a8156ce25"} Feb 27 17:01:43 crc kubenswrapper[4814]: I0227 17:01:43.271882 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d8bvb" podStartSLOduration=2.80817627 podStartE2EDuration="5.271862249s" podCreationTimestamp="2026-02-27 17:01:38 +0000 UTC" firstStartedPulling="2026-02-27 17:01:40.187293837 +0000 UTC m=+2312.639918667" lastFinishedPulling="2026-02-27 17:01:42.650979776 +0000 UTC m=+2315.103604646" observedRunningTime="2026-02-27 17:01:43.270969201 +0000 UTC m=+2315.723594101" watchObservedRunningTime="2026-02-27 17:01:43.271862249 +0000 UTC m=+2315.724487089" Feb 27 17:01:49 crc kubenswrapper[4814]: I0227 17:01:49.264487 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:49 crc kubenswrapper[4814]: I0227 17:01:49.265392 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:49 crc kubenswrapper[4814]: I0227 17:01:49.325144 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:49 crc kubenswrapper[4814]: I0227 17:01:49.399236 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:49 crc kubenswrapper[4814]: I0227 17:01:49.595036 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d8bvb"] Feb 27 17:01:51 crc kubenswrapper[4814]: I0227 17:01:51.338572 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d8bvb" podUID="8c5cf1bf-c514-4f94-9663-0760495dbe57" containerName="registry-server" containerID="cri-o://9cca9c0c274c438a18a0efe306f58b3e474ca0233accc1f0178d7f2a8156ce25" gracePeriod=2 Feb 27 17:01:51 crc kubenswrapper[4814]: I0227 17:01:51.821328 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:51 crc kubenswrapper[4814]: I0227 17:01:51.888752 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c5cf1bf-c514-4f94-9663-0760495dbe57-catalog-content\") pod \"8c5cf1bf-c514-4f94-9663-0760495dbe57\" (UID: \"8c5cf1bf-c514-4f94-9663-0760495dbe57\") " Feb 27 17:01:51 crc kubenswrapper[4814]: I0227 17:01:51.889811 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72tsq\" (UniqueName: \"kubernetes.io/projected/8c5cf1bf-c514-4f94-9663-0760495dbe57-kube-api-access-72tsq\") pod \"8c5cf1bf-c514-4f94-9663-0760495dbe57\" (UID: \"8c5cf1bf-c514-4f94-9663-0760495dbe57\") " Feb 27 17:01:51 crc kubenswrapper[4814]: I0227 17:01:51.889887 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c5cf1bf-c514-4f94-9663-0760495dbe57-utilities\") pod \"8c5cf1bf-c514-4f94-9663-0760495dbe57\" (UID: \"8c5cf1bf-c514-4f94-9663-0760495dbe57\") " Feb 27 17:01:51 crc kubenswrapper[4814]: I0227 17:01:51.892124 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c5cf1bf-c514-4f94-9663-0760495dbe57-utilities" (OuterVolumeSpecName: "utilities") pod "8c5cf1bf-c514-4f94-9663-0760495dbe57" (UID: "8c5cf1bf-c514-4f94-9663-0760495dbe57"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:01:51 crc kubenswrapper[4814]: I0227 17:01:51.901140 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c5cf1bf-c514-4f94-9663-0760495dbe57-kube-api-access-72tsq" (OuterVolumeSpecName: "kube-api-access-72tsq") pod "8c5cf1bf-c514-4f94-9663-0760495dbe57" (UID: "8c5cf1bf-c514-4f94-9663-0760495dbe57"). InnerVolumeSpecName "kube-api-access-72tsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:01:51 crc kubenswrapper[4814]: I0227 17:01:51.974635 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c5cf1bf-c514-4f94-9663-0760495dbe57-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c5cf1bf-c514-4f94-9663-0760495dbe57" (UID: "8c5cf1bf-c514-4f94-9663-0760495dbe57"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:01:51 crc kubenswrapper[4814]: I0227 17:01:51.992571 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72tsq\" (UniqueName: \"kubernetes.io/projected/8c5cf1bf-c514-4f94-9663-0760495dbe57-kube-api-access-72tsq\") on node \"crc\" DevicePath \"\"" Feb 27 17:01:51 crc kubenswrapper[4814]: I0227 17:01:51.992605 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c5cf1bf-c514-4f94-9663-0760495dbe57-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:01:51 crc kubenswrapper[4814]: I0227 17:01:51.992615 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c5cf1bf-c514-4f94-9663-0760495dbe57-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:01:52 crc kubenswrapper[4814]: I0227 17:01:52.350489 4814 generic.go:334] "Generic (PLEG): container finished" podID="8c5cf1bf-c514-4f94-9663-0760495dbe57" containerID="9cca9c0c274c438a18a0efe306f58b3e474ca0233accc1f0178d7f2a8156ce25" exitCode=0 Feb 27 17:01:52 crc kubenswrapper[4814]: I0227 17:01:52.350537 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8bvb" event={"ID":"8c5cf1bf-c514-4f94-9663-0760495dbe57","Type":"ContainerDied","Data":"9cca9c0c274c438a18a0efe306f58b3e474ca0233accc1f0178d7f2a8156ce25"} Feb 27 17:01:52 crc kubenswrapper[4814]: I0227 17:01:52.350567 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8bvb" event={"ID":"8c5cf1bf-c514-4f94-9663-0760495dbe57","Type":"ContainerDied","Data":"ed74c73f9dfc2d1594f695d7a09b9b861ef4f31a3ae6df280103ac46caf544d7"} Feb 27 17:01:52 crc kubenswrapper[4814]: I0227 17:01:52.350587 4814 scope.go:117] "RemoveContainer" containerID="9cca9c0c274c438a18a0efe306f58b3e474ca0233accc1f0178d7f2a8156ce25" Feb 27 17:01:52 crc kubenswrapper[4814]: I0227 17:01:52.351746 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d8bvb" Feb 27 17:01:52 crc kubenswrapper[4814]: I0227 17:01:52.378458 4814 scope.go:117] "RemoveContainer" containerID="458162105fbd39017477fe029904eb22e30723ea07aa9050b427a049342ca3c0" Feb 27 17:01:52 crc kubenswrapper[4814]: I0227 17:01:52.392502 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d8bvb"] Feb 27 17:01:52 crc kubenswrapper[4814]: I0227 17:01:52.401940 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d8bvb"] Feb 27 17:01:52 crc kubenswrapper[4814]: I0227 17:01:52.421422 4814 scope.go:117] "RemoveContainer" containerID="44fdcb5754c8fbf6f34ddc1c97287daf5786f9a2c5763f1e50954837724846eb" Feb 27 17:01:52 crc kubenswrapper[4814]: I0227 17:01:52.466465 4814 scope.go:117] "RemoveContainer" containerID="9cca9c0c274c438a18a0efe306f58b3e474ca0233accc1f0178d7f2a8156ce25" Feb 27 17:01:52 crc kubenswrapper[4814]: E0227 17:01:52.466969 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cca9c0c274c438a18a0efe306f58b3e474ca0233accc1f0178d7f2a8156ce25\": container with ID starting with 9cca9c0c274c438a18a0efe306f58b3e474ca0233accc1f0178d7f2a8156ce25 not found: ID does not exist" containerID="9cca9c0c274c438a18a0efe306f58b3e474ca0233accc1f0178d7f2a8156ce25" Feb 27 17:01:52 crc kubenswrapper[4814]: I0227 17:01:52.467022 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cca9c0c274c438a18a0efe306f58b3e474ca0233accc1f0178d7f2a8156ce25"} err="failed to get container status \"9cca9c0c274c438a18a0efe306f58b3e474ca0233accc1f0178d7f2a8156ce25\": rpc error: code = NotFound desc = could not find container \"9cca9c0c274c438a18a0efe306f58b3e474ca0233accc1f0178d7f2a8156ce25\": container with ID starting with 9cca9c0c274c438a18a0efe306f58b3e474ca0233accc1f0178d7f2a8156ce25 not found: ID does not exist" Feb 27 17:01:52 crc kubenswrapper[4814]: I0227 17:01:52.467055 4814 scope.go:117] "RemoveContainer" containerID="458162105fbd39017477fe029904eb22e30723ea07aa9050b427a049342ca3c0" Feb 27 17:01:52 crc kubenswrapper[4814]: E0227 17:01:52.467459 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"458162105fbd39017477fe029904eb22e30723ea07aa9050b427a049342ca3c0\": container with ID starting with 458162105fbd39017477fe029904eb22e30723ea07aa9050b427a049342ca3c0 not found: ID does not exist" containerID="458162105fbd39017477fe029904eb22e30723ea07aa9050b427a049342ca3c0" Feb 27 17:01:52 crc kubenswrapper[4814]: I0227 17:01:52.467500 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"458162105fbd39017477fe029904eb22e30723ea07aa9050b427a049342ca3c0"} err="failed to get container status \"458162105fbd39017477fe029904eb22e30723ea07aa9050b427a049342ca3c0\": rpc error: code = NotFound desc = could not find container \"458162105fbd39017477fe029904eb22e30723ea07aa9050b427a049342ca3c0\": container with ID starting with 458162105fbd39017477fe029904eb22e30723ea07aa9050b427a049342ca3c0 not found: ID does not exist" Feb 27 17:01:52 crc kubenswrapper[4814]: I0227 17:01:52.467524 4814 scope.go:117] "RemoveContainer" containerID="44fdcb5754c8fbf6f34ddc1c97287daf5786f9a2c5763f1e50954837724846eb" Feb 27 17:01:52 crc kubenswrapper[4814]: E0227 17:01:52.467834 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44fdcb5754c8fbf6f34ddc1c97287daf5786f9a2c5763f1e50954837724846eb\": container with ID starting with 44fdcb5754c8fbf6f34ddc1c97287daf5786f9a2c5763f1e50954837724846eb not found: ID does not exist" containerID="44fdcb5754c8fbf6f34ddc1c97287daf5786f9a2c5763f1e50954837724846eb" Feb 27 17:01:52 crc kubenswrapper[4814]: I0227 17:01:52.467877 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44fdcb5754c8fbf6f34ddc1c97287daf5786f9a2c5763f1e50954837724846eb"} err="failed to get container status \"44fdcb5754c8fbf6f34ddc1c97287daf5786f9a2c5763f1e50954837724846eb\": rpc error: code = NotFound desc = could not find container \"44fdcb5754c8fbf6f34ddc1c97287daf5786f9a2c5763f1e50954837724846eb\": container with ID starting with 44fdcb5754c8fbf6f34ddc1c97287daf5786f9a2c5763f1e50954837724846eb not found: ID does not exist" Feb 27 17:01:52 crc kubenswrapper[4814]: I0227 17:01:52.501157 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c5cf1bf-c514-4f94-9663-0760495dbe57" path="/var/lib/kubelet/pods/8c5cf1bf-c514-4f94-9663-0760495dbe57/volumes" Feb 27 17:02:00 crc kubenswrapper[4814]: I0227 17:02:00.170140 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536862-b2l7p"] Feb 27 17:02:00 crc kubenswrapper[4814]: E0227 17:02:00.172503 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c5cf1bf-c514-4f94-9663-0760495dbe57" containerName="extract-utilities" Feb 27 17:02:00 crc kubenswrapper[4814]: I0227 17:02:00.172579 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c5cf1bf-c514-4f94-9663-0760495dbe57" containerName="extract-utilities" Feb 27 17:02:00 crc kubenswrapper[4814]: E0227 17:02:00.172645 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c5cf1bf-c514-4f94-9663-0760495dbe57" containerName="registry-server" Feb 27 17:02:00 crc kubenswrapper[4814]: I0227 17:02:00.172661 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c5cf1bf-c514-4f94-9663-0760495dbe57" containerName="registry-server" Feb 27 17:02:00 crc kubenswrapper[4814]: E0227 17:02:00.172795 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c5cf1bf-c514-4f94-9663-0760495dbe57" containerName="extract-content" Feb 27 17:02:00 crc kubenswrapper[4814]: I0227 17:02:00.172865 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c5cf1bf-c514-4f94-9663-0760495dbe57" containerName="extract-content" Feb 27 17:02:00 crc kubenswrapper[4814]: I0227 17:02:00.173572 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c5cf1bf-c514-4f94-9663-0760495dbe57" containerName="registry-server" Feb 27 17:02:00 crc kubenswrapper[4814]: I0227 17:02:00.174762 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536862-b2l7p" Feb 27 17:02:00 crc kubenswrapper[4814]: I0227 17:02:00.178387 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:02:00 crc kubenswrapper[4814]: I0227 17:02:00.178439 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:02:00 crc kubenswrapper[4814]: I0227 17:02:00.179728 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:02:00 crc kubenswrapper[4814]: I0227 17:02:00.188568 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536862-b2l7p"] Feb 27 17:02:00 crc kubenswrapper[4814]: I0227 17:02:00.273048 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gljnd\" (UniqueName: \"kubernetes.io/projected/04d54354-19e5-4d8e-9899-eeb32b752e05-kube-api-access-gljnd\") pod \"auto-csr-approver-29536862-b2l7p\" (UID: \"04d54354-19e5-4d8e-9899-eeb32b752e05\") " pod="openshift-infra/auto-csr-approver-29536862-b2l7p" Feb 27 17:02:00 crc kubenswrapper[4814]: I0227 17:02:00.376227 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gljnd\" (UniqueName: \"kubernetes.io/projected/04d54354-19e5-4d8e-9899-eeb32b752e05-kube-api-access-gljnd\") pod \"auto-csr-approver-29536862-b2l7p\" (UID: \"04d54354-19e5-4d8e-9899-eeb32b752e05\") " pod="openshift-infra/auto-csr-approver-29536862-b2l7p" Feb 27 17:02:00 crc kubenswrapper[4814]: I0227 17:02:00.402171 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gljnd\" (UniqueName: \"kubernetes.io/projected/04d54354-19e5-4d8e-9899-eeb32b752e05-kube-api-access-gljnd\") pod \"auto-csr-approver-29536862-b2l7p\" (UID: \"04d54354-19e5-4d8e-9899-eeb32b752e05\") " pod="openshift-infra/auto-csr-approver-29536862-b2l7p" Feb 27 17:02:00 crc kubenswrapper[4814]: I0227 17:02:00.521332 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536862-b2l7p" Feb 27 17:02:00 crc kubenswrapper[4814]: I0227 17:02:00.977959 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536862-b2l7p"] Feb 27 17:02:00 crc kubenswrapper[4814]: W0227 17:02:00.983977 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04d54354_19e5_4d8e_9899_eeb32b752e05.slice/crio-3d4599f2a4d32303e4a74224320cfab6f09b1319c275bee7c18b57ab4c63c4b2 WatchSource:0}: Error finding container 3d4599f2a4d32303e4a74224320cfab6f09b1319c275bee7c18b57ab4c63c4b2: Status 404 returned error can't find the container with id 3d4599f2a4d32303e4a74224320cfab6f09b1319c275bee7c18b57ab4c63c4b2 Feb 27 17:02:01 crc kubenswrapper[4814]: I0227 17:02:01.457834 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536862-b2l7p" event={"ID":"04d54354-19e5-4d8e-9899-eeb32b752e05","Type":"ContainerStarted","Data":"3d4599f2a4d32303e4a74224320cfab6f09b1319c275bee7c18b57ab4c63c4b2"} Feb 27 17:02:03 crc kubenswrapper[4814]: I0227 17:02:03.479647 4814 generic.go:334] "Generic (PLEG): container finished" podID="04d54354-19e5-4d8e-9899-eeb32b752e05" containerID="0466b5febadec44c8ae866fef3e97d7cd950a179d4ff5a4bdebbf51f9b4a7970" exitCode=0 Feb 27 17:02:03 crc kubenswrapper[4814]: I0227 17:02:03.479731 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536862-b2l7p" event={"ID":"04d54354-19e5-4d8e-9899-eeb32b752e05","Type":"ContainerDied","Data":"0466b5febadec44c8ae866fef3e97d7cd950a179d4ff5a4bdebbf51f9b4a7970"} Feb 27 17:02:04 crc kubenswrapper[4814]: I0227 17:02:04.886068 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536862-b2l7p" Feb 27 17:02:05 crc kubenswrapper[4814]: I0227 17:02:05.001421 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gljnd\" (UniqueName: \"kubernetes.io/projected/04d54354-19e5-4d8e-9899-eeb32b752e05-kube-api-access-gljnd\") pod \"04d54354-19e5-4d8e-9899-eeb32b752e05\" (UID: \"04d54354-19e5-4d8e-9899-eeb32b752e05\") " Feb 27 17:02:05 crc kubenswrapper[4814]: I0227 17:02:05.009189 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04d54354-19e5-4d8e-9899-eeb32b752e05-kube-api-access-gljnd" (OuterVolumeSpecName: "kube-api-access-gljnd") pod "04d54354-19e5-4d8e-9899-eeb32b752e05" (UID: "04d54354-19e5-4d8e-9899-eeb32b752e05"). InnerVolumeSpecName "kube-api-access-gljnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:05 crc kubenswrapper[4814]: I0227 17:02:05.103453 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gljnd\" (UniqueName: \"kubernetes.io/projected/04d54354-19e5-4d8e-9899-eeb32b752e05-kube-api-access-gljnd\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:05 crc kubenswrapper[4814]: I0227 17:02:05.505737 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536862-b2l7p" event={"ID":"04d54354-19e5-4d8e-9899-eeb32b752e05","Type":"ContainerDied","Data":"3d4599f2a4d32303e4a74224320cfab6f09b1319c275bee7c18b57ab4c63c4b2"} Feb 27 17:02:05 crc kubenswrapper[4814]: I0227 17:02:05.506074 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d4599f2a4d32303e4a74224320cfab6f09b1319c275bee7c18b57ab4c63c4b2" Feb 27 17:02:05 crc kubenswrapper[4814]: I0227 17:02:05.505819 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536862-b2l7p" Feb 27 17:02:05 crc kubenswrapper[4814]: I0227 17:02:05.954989 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536856-9mgrh"] Feb 27 17:02:05 crc kubenswrapper[4814]: I0227 17:02:05.964919 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536856-9mgrh"] Feb 27 17:02:06 crc kubenswrapper[4814]: I0227 17:02:06.501478 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c9122d2-7427-4dd8-974d-38cfce5c3cad" path="/var/lib/kubelet/pods/3c9122d2-7427-4dd8-974d-38cfce5c3cad/volumes" Feb 27 17:02:57 crc kubenswrapper[4814]: I0227 17:02:57.048474 4814 scope.go:117] "RemoveContainer" containerID="3e6fd5aed07cd58e1536aaf9534be94f5bebd7fcf616b732cb96128307fe4245" Feb 27 17:03:52 crc kubenswrapper[4814]: I0227 17:03:52.902134 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:03:52 crc kubenswrapper[4814]: I0227 17:03:52.902698 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:04:00 crc kubenswrapper[4814]: I0227 17:04:00.160382 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536864-w6vcs"] Feb 27 17:04:00 crc kubenswrapper[4814]: E0227 17:04:00.161804 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d54354-19e5-4d8e-9899-eeb32b752e05" containerName="oc" Feb 27 17:04:00 crc kubenswrapper[4814]: I0227 17:04:00.161831 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d54354-19e5-4d8e-9899-eeb32b752e05" containerName="oc" Feb 27 17:04:00 crc kubenswrapper[4814]: I0227 17:04:00.162210 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="04d54354-19e5-4d8e-9899-eeb32b752e05" containerName="oc" Feb 27 17:04:00 crc kubenswrapper[4814]: I0227 17:04:00.163318 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536864-w6vcs" Feb 27 17:04:00 crc kubenswrapper[4814]: I0227 17:04:00.167208 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:04:00 crc kubenswrapper[4814]: I0227 17:04:00.167422 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:04:00 crc kubenswrapper[4814]: I0227 17:04:00.167516 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:04:00 crc kubenswrapper[4814]: I0227 17:04:00.170391 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536864-w6vcs"] Feb 27 17:04:00 crc kubenswrapper[4814]: I0227 17:04:00.205687 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8h4m\" (UniqueName: \"kubernetes.io/projected/19875f7e-5900-4480-af38-cb7e970b67d2-kube-api-access-c8h4m\") pod \"auto-csr-approver-29536864-w6vcs\" (UID: \"19875f7e-5900-4480-af38-cb7e970b67d2\") " pod="openshift-infra/auto-csr-approver-29536864-w6vcs" Feb 27 17:04:00 crc kubenswrapper[4814]: I0227 17:04:00.308483 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8h4m\" (UniqueName: \"kubernetes.io/projected/19875f7e-5900-4480-af38-cb7e970b67d2-kube-api-access-c8h4m\") pod \"auto-csr-approver-29536864-w6vcs\" (UID: \"19875f7e-5900-4480-af38-cb7e970b67d2\") " pod="openshift-infra/auto-csr-approver-29536864-w6vcs" Feb 27 17:04:00 crc kubenswrapper[4814]: I0227 17:04:00.331384 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8h4m\" (UniqueName: \"kubernetes.io/projected/19875f7e-5900-4480-af38-cb7e970b67d2-kube-api-access-c8h4m\") pod \"auto-csr-approver-29536864-w6vcs\" (UID: \"19875f7e-5900-4480-af38-cb7e970b67d2\") " pod="openshift-infra/auto-csr-approver-29536864-w6vcs" Feb 27 17:04:00 crc kubenswrapper[4814]: I0227 17:04:00.505517 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536864-w6vcs" Feb 27 17:04:00 crc kubenswrapper[4814]: I0227 17:04:00.904140 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536864-w6vcs"] Feb 27 17:04:00 crc kubenswrapper[4814]: I0227 17:04:00.912621 4814 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:04:01 crc kubenswrapper[4814]: I0227 17:04:01.883991 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536864-w6vcs" event={"ID":"19875f7e-5900-4480-af38-cb7e970b67d2","Type":"ContainerStarted","Data":"c1b92be5e6b9f99a25fd9c9cbb8e761bcb25f57e5a564bfe834cfef1a87e7376"} Feb 27 17:04:02 crc kubenswrapper[4814]: I0227 17:04:02.895806 4814 generic.go:334] "Generic (PLEG): container finished" podID="19875f7e-5900-4480-af38-cb7e970b67d2" containerID="9b7af82cc2107f82111d248255ef0d73a7e8d5ec24fb6ce2e86535ed73642963" exitCode=0 Feb 27 17:04:02 crc kubenswrapper[4814]: I0227 17:04:02.895860 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536864-w6vcs" event={"ID":"19875f7e-5900-4480-af38-cb7e970b67d2","Type":"ContainerDied","Data":"9b7af82cc2107f82111d248255ef0d73a7e8d5ec24fb6ce2e86535ed73642963"} Feb 27 17:04:04 crc kubenswrapper[4814]: I0227 17:04:04.280852 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536864-w6vcs" Feb 27 17:04:04 crc kubenswrapper[4814]: I0227 17:04:04.407957 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8h4m\" (UniqueName: \"kubernetes.io/projected/19875f7e-5900-4480-af38-cb7e970b67d2-kube-api-access-c8h4m\") pod \"19875f7e-5900-4480-af38-cb7e970b67d2\" (UID: \"19875f7e-5900-4480-af38-cb7e970b67d2\") " Feb 27 17:04:04 crc kubenswrapper[4814]: I0227 17:04:04.413689 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19875f7e-5900-4480-af38-cb7e970b67d2-kube-api-access-c8h4m" (OuterVolumeSpecName: "kube-api-access-c8h4m") pod "19875f7e-5900-4480-af38-cb7e970b67d2" (UID: "19875f7e-5900-4480-af38-cb7e970b67d2"). InnerVolumeSpecName "kube-api-access-c8h4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:04:04 crc kubenswrapper[4814]: I0227 17:04:04.510087 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8h4m\" (UniqueName: \"kubernetes.io/projected/19875f7e-5900-4480-af38-cb7e970b67d2-kube-api-access-c8h4m\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:04 crc kubenswrapper[4814]: I0227 17:04:04.922008 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536864-w6vcs" event={"ID":"19875f7e-5900-4480-af38-cb7e970b67d2","Type":"ContainerDied","Data":"c1b92be5e6b9f99a25fd9c9cbb8e761bcb25f57e5a564bfe834cfef1a87e7376"} Feb 27 17:04:04 crc kubenswrapper[4814]: I0227 17:04:04.922424 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1b92be5e6b9f99a25fd9c9cbb8e761bcb25f57e5a564bfe834cfef1a87e7376" Feb 27 17:04:04 crc kubenswrapper[4814]: I0227 17:04:04.922100 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536864-w6vcs" Feb 27 17:04:05 crc kubenswrapper[4814]: I0227 17:04:05.378151 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536858-z4w2z"] Feb 27 17:04:05 crc kubenswrapper[4814]: I0227 17:04:05.393747 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536858-z4w2z"] Feb 27 17:04:06 crc kubenswrapper[4814]: I0227 17:04:06.505968 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0060b006-c2cf-4a60-87b3-5c9f644d0ad9" path="/var/lib/kubelet/pods/0060b006-c2cf-4a60-87b3-5c9f644d0ad9/volumes" Feb 27 17:04:13 crc kubenswrapper[4814]: I0227 17:04:13.000785 4814 generic.go:334] "Generic (PLEG): container finished" podID="c96b1d1f-6478-4eea-8413-a0657f19c6e7" containerID="35730396da47ff0b822a5f1a5840f41d14871ea95fb1f4cc78ff307f88a3d6ed" exitCode=0 Feb 27 17:04:13 crc kubenswrapper[4814]: I0227 17:04:13.000889 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" event={"ID":"c96b1d1f-6478-4eea-8413-a0657f19c6e7","Type":"ContainerDied","Data":"35730396da47ff0b822a5f1a5840f41d14871ea95fb1f4cc78ff307f88a3d6ed"} Feb 27 17:04:14 crc kubenswrapper[4814]: I0227 17:04:14.540445 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 17:04:14 crc kubenswrapper[4814]: I0227 17:04:14.648399 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn7vx\" (UniqueName: \"kubernetes.io/projected/c96b1d1f-6478-4eea-8413-a0657f19c6e7-kube-api-access-zn7vx\") pod \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " Feb 27 17:04:14 crc kubenswrapper[4814]: I0227 17:04:14.648482 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-ssh-key-openstack-edpm-ipam\") pod \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " Feb 27 17:04:14 crc kubenswrapper[4814]: I0227 17:04:14.648514 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-inventory\") pod \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " Feb 27 17:04:14 crc kubenswrapper[4814]: I0227 17:04:14.648653 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-libvirt-combined-ca-bundle\") pod \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " Feb 27 17:04:14 crc kubenswrapper[4814]: I0227 17:04:14.648683 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-libvirt-secret-0\") pod \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\" (UID: \"c96b1d1f-6478-4eea-8413-a0657f19c6e7\") " Feb 27 17:04:14 crc kubenswrapper[4814]: I0227 17:04:14.653345 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c96b1d1f-6478-4eea-8413-a0657f19c6e7" (UID: "c96b1d1f-6478-4eea-8413-a0657f19c6e7"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:04:14 crc kubenswrapper[4814]: I0227 17:04:14.666521 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c96b1d1f-6478-4eea-8413-a0657f19c6e7-kube-api-access-zn7vx" (OuterVolumeSpecName: "kube-api-access-zn7vx") pod "c96b1d1f-6478-4eea-8413-a0657f19c6e7" (UID: "c96b1d1f-6478-4eea-8413-a0657f19c6e7"). InnerVolumeSpecName "kube-api-access-zn7vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:04:14 crc kubenswrapper[4814]: I0227 17:04:14.686276 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "c96b1d1f-6478-4eea-8413-a0657f19c6e7" (UID: "c96b1d1f-6478-4eea-8413-a0657f19c6e7"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:04:14 crc kubenswrapper[4814]: I0227 17:04:14.689781 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c96b1d1f-6478-4eea-8413-a0657f19c6e7" (UID: "c96b1d1f-6478-4eea-8413-a0657f19c6e7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:04:14 crc kubenswrapper[4814]: I0227 17:04:14.696441 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-inventory" (OuterVolumeSpecName: "inventory") pod "c96b1d1f-6478-4eea-8413-a0657f19c6e7" (UID: "c96b1d1f-6478-4eea-8413-a0657f19c6e7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:04:14 crc kubenswrapper[4814]: I0227 17:04:14.762629 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn7vx\" (UniqueName: \"kubernetes.io/projected/c96b1d1f-6478-4eea-8413-a0657f19c6e7-kube-api-access-zn7vx\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:14 crc kubenswrapper[4814]: I0227 17:04:14.762695 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:14 crc kubenswrapper[4814]: I0227 17:04:14.762712 4814 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:14 crc kubenswrapper[4814]: I0227 17:04:14.762723 4814 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:14 crc kubenswrapper[4814]: I0227 17:04:14.762787 4814 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c96b1d1f-6478-4eea-8413-a0657f19c6e7-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.027093 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" event={"ID":"c96b1d1f-6478-4eea-8413-a0657f19c6e7","Type":"ContainerDied","Data":"b7dc90de91b166e4b7980aa4047accbea01899b440981816ef3760455e011130"} Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.027145 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7dc90de91b166e4b7980aa4047accbea01899b440981816ef3760455e011130" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.027219 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.167422 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m"] Feb 27 17:04:15 crc kubenswrapper[4814]: E0227 17:04:15.167988 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19875f7e-5900-4480-af38-cb7e970b67d2" containerName="oc" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.168011 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="19875f7e-5900-4480-af38-cb7e970b67d2" containerName="oc" Feb 27 17:04:15 crc kubenswrapper[4814]: E0227 17:04:15.168044 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c96b1d1f-6478-4eea-8413-a0657f19c6e7" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.168054 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="c96b1d1f-6478-4eea-8413-a0657f19c6e7" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.168347 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="c96b1d1f-6478-4eea-8413-a0657f19c6e7" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.168387 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="19875f7e-5900-4480-af38-cb7e970b67d2" containerName="oc" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.169220 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.174629 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.176175 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.176184 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.176237 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.176181 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.176729 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.177382 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.184285 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m"] Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.271684 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.271731 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.271987 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.272036 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.272078 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.272147 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.272184 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.272211 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.272323 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.272351 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.272402 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd2tj\" (UniqueName: \"kubernetes.io/projected/2f3a5797-c53e-4195-8e15-f2ba2016c410-kube-api-access-dd2tj\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.374058 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.374137 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.374310 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.374356 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.374495 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.374552 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.374598 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.374638 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.374718 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.374771 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.374825 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd2tj\" (UniqueName: \"kubernetes.io/projected/2f3a5797-c53e-4195-8e15-f2ba2016c410-kube-api-access-dd2tj\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.375760 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.381357 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.382145 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.382387 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.382445 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.385247 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.385504 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.386794 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.387172 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.393144 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.406536 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd2tj\" (UniqueName: \"kubernetes.io/projected/2f3a5797-c53e-4195-8e15-f2ba2016c410-kube-api-access-dd2tj\") pod \"nova-edpm-deployment-openstack-edpm-ipam-48n4m\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:15 crc kubenswrapper[4814]: I0227 17:04:15.498012 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:04:16 crc kubenswrapper[4814]: I0227 17:04:16.143997 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m"] Feb 27 17:04:17 crc kubenswrapper[4814]: I0227 17:04:17.048898 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" event={"ID":"2f3a5797-c53e-4195-8e15-f2ba2016c410","Type":"ContainerStarted","Data":"9f0bc6db4ababccc6c692011a3e2a5c3ad1262beea05e57298962c41feabead8"} Feb 27 17:04:17 crc kubenswrapper[4814]: I0227 17:04:17.049366 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" event={"ID":"2f3a5797-c53e-4195-8e15-f2ba2016c410","Type":"ContainerStarted","Data":"8c8d7c3eda3e5e920f27fc63d41f1e43c8224047bd2abdce7cea2ae46c88e4bd"} Feb 27 17:04:17 crc kubenswrapper[4814]: I0227 17:04:17.083507 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" podStartSLOduration=1.671545289 podStartE2EDuration="2.083487618s" podCreationTimestamp="2026-02-27 17:04:15 +0000 UTC" firstStartedPulling="2026-02-27 17:04:16.155461807 +0000 UTC m=+2468.608086637" lastFinishedPulling="2026-02-27 17:04:16.567404136 +0000 UTC m=+2469.020028966" observedRunningTime="2026-02-27 17:04:17.077171271 +0000 UTC m=+2469.529796121" watchObservedRunningTime="2026-02-27 17:04:17.083487618 +0000 UTC m=+2469.536112458" Feb 27 17:04:22 crc kubenswrapper[4814]: I0227 17:04:22.902541 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:04:22 crc kubenswrapper[4814]: I0227 17:04:22.903173 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:04:52 crc kubenswrapper[4814]: I0227 17:04:52.902521 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:04:52 crc kubenswrapper[4814]: I0227 17:04:52.903336 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:04:52 crc kubenswrapper[4814]: I0227 17:04:52.903407 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 17:04:52 crc kubenswrapper[4814]: I0227 17:04:52.904443 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a"} pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:04:52 crc kubenswrapper[4814]: I0227 17:04:52.904536 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" containerID="cri-o://8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" gracePeriod=600 Feb 27 17:04:53 crc kubenswrapper[4814]: E0227 17:04:53.033030 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:04:53 crc kubenswrapper[4814]: I0227 17:04:53.439430 4814 generic.go:334] "Generic (PLEG): container finished" podID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" exitCode=0 Feb 27 17:04:53 crc kubenswrapper[4814]: I0227 17:04:53.439475 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerDied","Data":"8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a"} Feb 27 17:04:53 crc kubenswrapper[4814]: I0227 17:04:53.439512 4814 scope.go:117] "RemoveContainer" containerID="eb9a72219e9b318e493f16a1ab078d7373596a6f5f4b8c4c5a593b8ea470dc7c" Feb 27 17:04:53 crc kubenswrapper[4814]: I0227 17:04:53.441700 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:04:53 crc kubenswrapper[4814]: E0227 17:04:53.441967 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:04:57 crc kubenswrapper[4814]: I0227 17:04:57.153543 4814 scope.go:117] "RemoveContainer" containerID="adbef7afb182a7c61ac9f9b627652bbae787ceee9e64bb40aa473679d4df7cf4" Feb 27 17:05:08 crc kubenswrapper[4814]: I0227 17:05:08.496284 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:05:08 crc kubenswrapper[4814]: E0227 17:05:08.497126 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:05:19 crc kubenswrapper[4814]: I0227 17:05:19.488464 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:05:19 crc kubenswrapper[4814]: E0227 17:05:19.489760 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:05:20 crc kubenswrapper[4814]: I0227 17:05:20.578317 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wknbx"] Feb 27 17:05:20 crc kubenswrapper[4814]: I0227 17:05:20.580620 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:20 crc kubenswrapper[4814]: I0227 17:05:20.586439 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wknbx"] Feb 27 17:05:20 crc kubenswrapper[4814]: I0227 17:05:20.665602 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rscn5\" (UniqueName: \"kubernetes.io/projected/cf651dba-501c-4359-89e6-1ba5c0b3547a-kube-api-access-rscn5\") pod \"community-operators-wknbx\" (UID: \"cf651dba-501c-4359-89e6-1ba5c0b3547a\") " pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:20 crc kubenswrapper[4814]: I0227 17:05:20.665742 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf651dba-501c-4359-89e6-1ba5c0b3547a-catalog-content\") pod \"community-operators-wknbx\" (UID: \"cf651dba-501c-4359-89e6-1ba5c0b3547a\") " pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:20 crc kubenswrapper[4814]: I0227 17:05:20.665786 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf651dba-501c-4359-89e6-1ba5c0b3547a-utilities\") pod \"community-operators-wknbx\" (UID: \"cf651dba-501c-4359-89e6-1ba5c0b3547a\") " pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:20 crc kubenswrapper[4814]: I0227 17:05:20.767498 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rscn5\" (UniqueName: \"kubernetes.io/projected/cf651dba-501c-4359-89e6-1ba5c0b3547a-kube-api-access-rscn5\") pod \"community-operators-wknbx\" (UID: \"cf651dba-501c-4359-89e6-1ba5c0b3547a\") " pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:20 crc kubenswrapper[4814]: I0227 17:05:20.767606 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf651dba-501c-4359-89e6-1ba5c0b3547a-catalog-content\") pod \"community-operators-wknbx\" (UID: \"cf651dba-501c-4359-89e6-1ba5c0b3547a\") " pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:20 crc kubenswrapper[4814]: I0227 17:05:20.767653 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf651dba-501c-4359-89e6-1ba5c0b3547a-utilities\") pod \"community-operators-wknbx\" (UID: \"cf651dba-501c-4359-89e6-1ba5c0b3547a\") " pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:20 crc kubenswrapper[4814]: I0227 17:05:20.768233 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf651dba-501c-4359-89e6-1ba5c0b3547a-utilities\") pod \"community-operators-wknbx\" (UID: \"cf651dba-501c-4359-89e6-1ba5c0b3547a\") " pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:20 crc kubenswrapper[4814]: I0227 17:05:20.768550 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf651dba-501c-4359-89e6-1ba5c0b3547a-catalog-content\") pod \"community-operators-wknbx\" (UID: \"cf651dba-501c-4359-89e6-1ba5c0b3547a\") " pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:20 crc kubenswrapper[4814]: I0227 17:05:20.791191 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rscn5\" (UniqueName: \"kubernetes.io/projected/cf651dba-501c-4359-89e6-1ba5c0b3547a-kube-api-access-rscn5\") pod \"community-operators-wknbx\" (UID: \"cf651dba-501c-4359-89e6-1ba5c0b3547a\") " pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:20 crc kubenswrapper[4814]: I0227 17:05:20.901826 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:21 crc kubenswrapper[4814]: W0227 17:05:21.520450 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf651dba_501c_4359_89e6_1ba5c0b3547a.slice/crio-631be7beac9549c9befb4745920501424bb96e5b1748caa09b044c3af1acdd1b WatchSource:0}: Error finding container 631be7beac9549c9befb4745920501424bb96e5b1748caa09b044c3af1acdd1b: Status 404 returned error can't find the container with id 631be7beac9549c9befb4745920501424bb96e5b1748caa09b044c3af1acdd1b Feb 27 17:05:21 crc kubenswrapper[4814]: I0227 17:05:21.526758 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wknbx"] Feb 27 17:05:21 crc kubenswrapper[4814]: I0227 17:05:21.817934 4814 generic.go:334] "Generic (PLEG): container finished" podID="cf651dba-501c-4359-89e6-1ba5c0b3547a" containerID="950df3c882cb0c0a81b3ab07c21286d6c32867cd40467fa8770dcba2c1c080a9" exitCode=0 Feb 27 17:05:21 crc kubenswrapper[4814]: I0227 17:05:21.817977 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wknbx" event={"ID":"cf651dba-501c-4359-89e6-1ba5c0b3547a","Type":"ContainerDied","Data":"950df3c882cb0c0a81b3ab07c21286d6c32867cd40467fa8770dcba2c1c080a9"} Feb 27 17:05:21 crc kubenswrapper[4814]: I0227 17:05:21.818004 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wknbx" event={"ID":"cf651dba-501c-4359-89e6-1ba5c0b3547a","Type":"ContainerStarted","Data":"631be7beac9549c9befb4745920501424bb96e5b1748caa09b044c3af1acdd1b"} Feb 27 17:05:22 crc kubenswrapper[4814]: I0227 17:05:22.829545 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wknbx" event={"ID":"cf651dba-501c-4359-89e6-1ba5c0b3547a","Type":"ContainerStarted","Data":"5a1ccf5b241c0889df77584a435c09bcdd92d21d1b58f0f0e4e551a18238f259"} Feb 27 17:05:23 crc kubenswrapper[4814]: I0227 17:05:23.844773 4814 generic.go:334] "Generic (PLEG): container finished" podID="cf651dba-501c-4359-89e6-1ba5c0b3547a" containerID="5a1ccf5b241c0889df77584a435c09bcdd92d21d1b58f0f0e4e551a18238f259" exitCode=0 Feb 27 17:05:23 crc kubenswrapper[4814]: I0227 17:05:23.844942 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wknbx" event={"ID":"cf651dba-501c-4359-89e6-1ba5c0b3547a","Type":"ContainerDied","Data":"5a1ccf5b241c0889df77584a435c09bcdd92d21d1b58f0f0e4e551a18238f259"} Feb 27 17:05:24 crc kubenswrapper[4814]: I0227 17:05:24.857518 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wknbx" event={"ID":"cf651dba-501c-4359-89e6-1ba5c0b3547a","Type":"ContainerStarted","Data":"6226c6c5aa5ecbd3fcaf6cbac62fc0c8c4bf8160726ff5483aa6ce22998f683a"} Feb 27 17:05:24 crc kubenswrapper[4814]: I0227 17:05:24.898413 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wknbx" podStartSLOduration=2.474657701 podStartE2EDuration="4.898386095s" podCreationTimestamp="2026-02-27 17:05:20 +0000 UTC" firstStartedPulling="2026-02-27 17:05:21.820155413 +0000 UTC m=+2534.272780243" lastFinishedPulling="2026-02-27 17:05:24.243883777 +0000 UTC m=+2536.696508637" observedRunningTime="2026-02-27 17:05:24.881643675 +0000 UTC m=+2537.334268515" watchObservedRunningTime="2026-02-27 17:05:24.898386095 +0000 UTC m=+2537.351010965" Feb 27 17:05:30 crc kubenswrapper[4814]: I0227 17:05:30.489816 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:05:30 crc kubenswrapper[4814]: E0227 17:05:30.490908 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:05:30 crc kubenswrapper[4814]: I0227 17:05:30.902975 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:30 crc kubenswrapper[4814]: I0227 17:05:30.903164 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:30 crc kubenswrapper[4814]: I0227 17:05:30.973658 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:31 crc kubenswrapper[4814]: I0227 17:05:31.043947 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:31 crc kubenswrapper[4814]: I0227 17:05:31.224940 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wknbx"] Feb 27 17:05:32 crc kubenswrapper[4814]: I0227 17:05:32.949792 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wknbx" podUID="cf651dba-501c-4359-89e6-1ba5c0b3547a" containerName="registry-server" containerID="cri-o://6226c6c5aa5ecbd3fcaf6cbac62fc0c8c4bf8160726ff5483aa6ce22998f683a" gracePeriod=2 Feb 27 17:05:33 crc kubenswrapper[4814]: I0227 17:05:33.391685 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:33 crc kubenswrapper[4814]: I0227 17:05:33.544405 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf651dba-501c-4359-89e6-1ba5c0b3547a-catalog-content\") pod \"cf651dba-501c-4359-89e6-1ba5c0b3547a\" (UID: \"cf651dba-501c-4359-89e6-1ba5c0b3547a\") " Feb 27 17:05:33 crc kubenswrapper[4814]: I0227 17:05:33.544539 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rscn5\" (UniqueName: \"kubernetes.io/projected/cf651dba-501c-4359-89e6-1ba5c0b3547a-kube-api-access-rscn5\") pod \"cf651dba-501c-4359-89e6-1ba5c0b3547a\" (UID: \"cf651dba-501c-4359-89e6-1ba5c0b3547a\") " Feb 27 17:05:33 crc kubenswrapper[4814]: I0227 17:05:33.544693 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf651dba-501c-4359-89e6-1ba5c0b3547a-utilities\") pod \"cf651dba-501c-4359-89e6-1ba5c0b3547a\" (UID: \"cf651dba-501c-4359-89e6-1ba5c0b3547a\") " Feb 27 17:05:33 crc kubenswrapper[4814]: I0227 17:05:33.546176 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf651dba-501c-4359-89e6-1ba5c0b3547a-utilities" (OuterVolumeSpecName: "utilities") pod "cf651dba-501c-4359-89e6-1ba5c0b3547a" (UID: "cf651dba-501c-4359-89e6-1ba5c0b3547a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:05:33 crc kubenswrapper[4814]: I0227 17:05:33.557595 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf651dba-501c-4359-89e6-1ba5c0b3547a-kube-api-access-rscn5" (OuterVolumeSpecName: "kube-api-access-rscn5") pod "cf651dba-501c-4359-89e6-1ba5c0b3547a" (UID: "cf651dba-501c-4359-89e6-1ba5c0b3547a"). InnerVolumeSpecName "kube-api-access-rscn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:05:33 crc kubenswrapper[4814]: I0227 17:05:33.596659 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf651dba-501c-4359-89e6-1ba5c0b3547a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cf651dba-501c-4359-89e6-1ba5c0b3547a" (UID: "cf651dba-501c-4359-89e6-1ba5c0b3547a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:05:33 crc kubenswrapper[4814]: I0227 17:05:33.647365 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf651dba-501c-4359-89e6-1ba5c0b3547a-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:33 crc kubenswrapper[4814]: I0227 17:05:33.647396 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf651dba-501c-4359-89e6-1ba5c0b3547a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:33 crc kubenswrapper[4814]: I0227 17:05:33.647406 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rscn5\" (UniqueName: \"kubernetes.io/projected/cf651dba-501c-4359-89e6-1ba5c0b3547a-kube-api-access-rscn5\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:33 crc kubenswrapper[4814]: I0227 17:05:33.965669 4814 generic.go:334] "Generic (PLEG): container finished" podID="cf651dba-501c-4359-89e6-1ba5c0b3547a" containerID="6226c6c5aa5ecbd3fcaf6cbac62fc0c8c4bf8160726ff5483aa6ce22998f683a" exitCode=0 Feb 27 17:05:33 crc kubenswrapper[4814]: I0227 17:05:33.965731 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wknbx" event={"ID":"cf651dba-501c-4359-89e6-1ba5c0b3547a","Type":"ContainerDied","Data":"6226c6c5aa5ecbd3fcaf6cbac62fc0c8c4bf8160726ff5483aa6ce22998f683a"} Feb 27 17:05:33 crc kubenswrapper[4814]: I0227 17:05:33.965764 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wknbx" Feb 27 17:05:33 crc kubenswrapper[4814]: I0227 17:05:33.967282 4814 scope.go:117] "RemoveContainer" containerID="6226c6c5aa5ecbd3fcaf6cbac62fc0c8c4bf8160726ff5483aa6ce22998f683a" Feb 27 17:05:33 crc kubenswrapper[4814]: I0227 17:05:33.967144 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wknbx" event={"ID":"cf651dba-501c-4359-89e6-1ba5c0b3547a","Type":"ContainerDied","Data":"631be7beac9549c9befb4745920501424bb96e5b1748caa09b044c3af1acdd1b"} Feb 27 17:05:33 crc kubenswrapper[4814]: I0227 17:05:33.994825 4814 scope.go:117] "RemoveContainer" containerID="5a1ccf5b241c0889df77584a435c09bcdd92d21d1b58f0f0e4e551a18238f259" Feb 27 17:05:34 crc kubenswrapper[4814]: I0227 17:05:34.019469 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wknbx"] Feb 27 17:05:34 crc kubenswrapper[4814]: I0227 17:05:34.030363 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wknbx"] Feb 27 17:05:34 crc kubenswrapper[4814]: I0227 17:05:34.039614 4814 scope.go:117] "RemoveContainer" containerID="950df3c882cb0c0a81b3ab07c21286d6c32867cd40467fa8770dcba2c1c080a9" Feb 27 17:05:34 crc kubenswrapper[4814]: I0227 17:05:34.094019 4814 scope.go:117] "RemoveContainer" containerID="6226c6c5aa5ecbd3fcaf6cbac62fc0c8c4bf8160726ff5483aa6ce22998f683a" Feb 27 17:05:34 crc kubenswrapper[4814]: E0227 17:05:34.095148 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6226c6c5aa5ecbd3fcaf6cbac62fc0c8c4bf8160726ff5483aa6ce22998f683a\": container with ID starting with 6226c6c5aa5ecbd3fcaf6cbac62fc0c8c4bf8160726ff5483aa6ce22998f683a not found: ID does not exist" containerID="6226c6c5aa5ecbd3fcaf6cbac62fc0c8c4bf8160726ff5483aa6ce22998f683a" Feb 27 17:05:34 crc kubenswrapper[4814]: I0227 17:05:34.095449 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6226c6c5aa5ecbd3fcaf6cbac62fc0c8c4bf8160726ff5483aa6ce22998f683a"} err="failed to get container status \"6226c6c5aa5ecbd3fcaf6cbac62fc0c8c4bf8160726ff5483aa6ce22998f683a\": rpc error: code = NotFound desc = could not find container \"6226c6c5aa5ecbd3fcaf6cbac62fc0c8c4bf8160726ff5483aa6ce22998f683a\": container with ID starting with 6226c6c5aa5ecbd3fcaf6cbac62fc0c8c4bf8160726ff5483aa6ce22998f683a not found: ID does not exist" Feb 27 17:05:34 crc kubenswrapper[4814]: I0227 17:05:34.095476 4814 scope.go:117] "RemoveContainer" containerID="5a1ccf5b241c0889df77584a435c09bcdd92d21d1b58f0f0e4e551a18238f259" Feb 27 17:05:34 crc kubenswrapper[4814]: E0227 17:05:34.095839 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a1ccf5b241c0889df77584a435c09bcdd92d21d1b58f0f0e4e551a18238f259\": container with ID starting with 5a1ccf5b241c0889df77584a435c09bcdd92d21d1b58f0f0e4e551a18238f259 not found: ID does not exist" containerID="5a1ccf5b241c0889df77584a435c09bcdd92d21d1b58f0f0e4e551a18238f259" Feb 27 17:05:34 crc kubenswrapper[4814]: I0227 17:05:34.095891 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a1ccf5b241c0889df77584a435c09bcdd92d21d1b58f0f0e4e551a18238f259"} err="failed to get container status \"5a1ccf5b241c0889df77584a435c09bcdd92d21d1b58f0f0e4e551a18238f259\": rpc error: code = NotFound desc = could not find container \"5a1ccf5b241c0889df77584a435c09bcdd92d21d1b58f0f0e4e551a18238f259\": container with ID starting with 5a1ccf5b241c0889df77584a435c09bcdd92d21d1b58f0f0e4e551a18238f259 not found: ID does not exist" Feb 27 17:05:34 crc kubenswrapper[4814]: I0227 17:05:34.095923 4814 scope.go:117] "RemoveContainer" containerID="950df3c882cb0c0a81b3ab07c21286d6c32867cd40467fa8770dcba2c1c080a9" Feb 27 17:05:34 crc kubenswrapper[4814]: E0227 17:05:34.096222 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"950df3c882cb0c0a81b3ab07c21286d6c32867cd40467fa8770dcba2c1c080a9\": container with ID starting with 950df3c882cb0c0a81b3ab07c21286d6c32867cd40467fa8770dcba2c1c080a9 not found: ID does not exist" containerID="950df3c882cb0c0a81b3ab07c21286d6c32867cd40467fa8770dcba2c1c080a9" Feb 27 17:05:34 crc kubenswrapper[4814]: I0227 17:05:34.096258 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"950df3c882cb0c0a81b3ab07c21286d6c32867cd40467fa8770dcba2c1c080a9"} err="failed to get container status \"950df3c882cb0c0a81b3ab07c21286d6c32867cd40467fa8770dcba2c1c080a9\": rpc error: code = NotFound desc = could not find container \"950df3c882cb0c0a81b3ab07c21286d6c32867cd40467fa8770dcba2c1c080a9\": container with ID starting with 950df3c882cb0c0a81b3ab07c21286d6c32867cd40467fa8770dcba2c1c080a9 not found: ID does not exist" Feb 27 17:05:34 crc kubenswrapper[4814]: I0227 17:05:34.503242 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf651dba-501c-4359-89e6-1ba5c0b3547a" path="/var/lib/kubelet/pods/cf651dba-501c-4359-89e6-1ba5c0b3547a/volumes" Feb 27 17:05:43 crc kubenswrapper[4814]: I0227 17:05:43.488752 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:05:43 crc kubenswrapper[4814]: E0227 17:05:43.489754 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:05:58 crc kubenswrapper[4814]: I0227 17:05:58.500687 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:05:58 crc kubenswrapper[4814]: E0227 17:05:58.501719 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:06:00 crc kubenswrapper[4814]: I0227 17:06:00.150056 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536866-qsflg"] Feb 27 17:06:00 crc kubenswrapper[4814]: E0227 17:06:00.151042 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf651dba-501c-4359-89e6-1ba5c0b3547a" containerName="extract-utilities" Feb 27 17:06:00 crc kubenswrapper[4814]: I0227 17:06:00.151064 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf651dba-501c-4359-89e6-1ba5c0b3547a" containerName="extract-utilities" Feb 27 17:06:00 crc kubenswrapper[4814]: E0227 17:06:00.151086 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf651dba-501c-4359-89e6-1ba5c0b3547a" containerName="extract-content" Feb 27 17:06:00 crc kubenswrapper[4814]: I0227 17:06:00.151099 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf651dba-501c-4359-89e6-1ba5c0b3547a" containerName="extract-content" Feb 27 17:06:00 crc kubenswrapper[4814]: E0227 17:06:00.151133 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf651dba-501c-4359-89e6-1ba5c0b3547a" containerName="registry-server" Feb 27 17:06:00 crc kubenswrapper[4814]: I0227 17:06:00.151147 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf651dba-501c-4359-89e6-1ba5c0b3547a" containerName="registry-server" Feb 27 17:06:00 crc kubenswrapper[4814]: I0227 17:06:00.151507 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf651dba-501c-4359-89e6-1ba5c0b3547a" containerName="registry-server" Feb 27 17:06:00 crc kubenswrapper[4814]: I0227 17:06:00.152560 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536866-qsflg" Feb 27 17:06:00 crc kubenswrapper[4814]: I0227 17:06:00.155858 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:06:00 crc kubenswrapper[4814]: I0227 17:06:00.156373 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:06:00 crc kubenswrapper[4814]: I0227 17:06:00.156491 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:06:00 crc kubenswrapper[4814]: I0227 17:06:00.161690 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536866-qsflg"] Feb 27 17:06:00 crc kubenswrapper[4814]: I0227 17:06:00.301389 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g85x\" (UniqueName: \"kubernetes.io/projected/b18ebba8-0085-45f5-b3da-fa039307426c-kube-api-access-5g85x\") pod \"auto-csr-approver-29536866-qsflg\" (UID: \"b18ebba8-0085-45f5-b3da-fa039307426c\") " pod="openshift-infra/auto-csr-approver-29536866-qsflg" Feb 27 17:06:00 crc kubenswrapper[4814]: I0227 17:06:00.403234 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g85x\" (UniqueName: \"kubernetes.io/projected/b18ebba8-0085-45f5-b3da-fa039307426c-kube-api-access-5g85x\") pod \"auto-csr-approver-29536866-qsflg\" (UID: \"b18ebba8-0085-45f5-b3da-fa039307426c\") " pod="openshift-infra/auto-csr-approver-29536866-qsflg" Feb 27 17:06:00 crc kubenswrapper[4814]: I0227 17:06:00.435976 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g85x\" (UniqueName: \"kubernetes.io/projected/b18ebba8-0085-45f5-b3da-fa039307426c-kube-api-access-5g85x\") pod \"auto-csr-approver-29536866-qsflg\" (UID: \"b18ebba8-0085-45f5-b3da-fa039307426c\") " pod="openshift-infra/auto-csr-approver-29536866-qsflg" Feb 27 17:06:00 crc kubenswrapper[4814]: I0227 17:06:00.481494 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536866-qsflg" Feb 27 17:06:01 crc kubenswrapper[4814]: I0227 17:06:01.528367 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536866-qsflg"] Feb 27 17:06:01 crc kubenswrapper[4814]: W0227 17:06:01.540600 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb18ebba8_0085_45f5_b3da_fa039307426c.slice/crio-ae96516856a2eea531cf43e68441e43737ec051aa7bf91c34a560c064253d12a WatchSource:0}: Error finding container ae96516856a2eea531cf43e68441e43737ec051aa7bf91c34a560c064253d12a: Status 404 returned error can't find the container with id ae96516856a2eea531cf43e68441e43737ec051aa7bf91c34a560c064253d12a Feb 27 17:06:02 crc kubenswrapper[4814]: I0227 17:06:02.321130 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536866-qsflg" event={"ID":"b18ebba8-0085-45f5-b3da-fa039307426c","Type":"ContainerStarted","Data":"ae96516856a2eea531cf43e68441e43737ec051aa7bf91c34a560c064253d12a"} Feb 27 17:06:03 crc kubenswrapper[4814]: I0227 17:06:03.333807 4814 generic.go:334] "Generic (PLEG): container finished" podID="b18ebba8-0085-45f5-b3da-fa039307426c" containerID="043a5d2eb1d7bba4a9977ca8ca21b9bab9f0b44e4688076178e6ca5a0ff4157d" exitCode=0 Feb 27 17:06:03 crc kubenswrapper[4814]: I0227 17:06:03.333891 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536866-qsflg" event={"ID":"b18ebba8-0085-45f5-b3da-fa039307426c","Type":"ContainerDied","Data":"043a5d2eb1d7bba4a9977ca8ca21b9bab9f0b44e4688076178e6ca5a0ff4157d"} Feb 27 17:06:04 crc kubenswrapper[4814]: I0227 17:06:04.754105 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536866-qsflg" Feb 27 17:06:04 crc kubenswrapper[4814]: I0227 17:06:04.909482 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5g85x\" (UniqueName: \"kubernetes.io/projected/b18ebba8-0085-45f5-b3da-fa039307426c-kube-api-access-5g85x\") pod \"b18ebba8-0085-45f5-b3da-fa039307426c\" (UID: \"b18ebba8-0085-45f5-b3da-fa039307426c\") " Feb 27 17:06:04 crc kubenswrapper[4814]: I0227 17:06:04.922485 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b18ebba8-0085-45f5-b3da-fa039307426c-kube-api-access-5g85x" (OuterVolumeSpecName: "kube-api-access-5g85x") pod "b18ebba8-0085-45f5-b3da-fa039307426c" (UID: "b18ebba8-0085-45f5-b3da-fa039307426c"). InnerVolumeSpecName "kube-api-access-5g85x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:06:05 crc kubenswrapper[4814]: I0227 17:06:05.011822 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5g85x\" (UniqueName: \"kubernetes.io/projected/b18ebba8-0085-45f5-b3da-fa039307426c-kube-api-access-5g85x\") on node \"crc\" DevicePath \"\"" Feb 27 17:06:05 crc kubenswrapper[4814]: I0227 17:06:05.357017 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536866-qsflg" event={"ID":"b18ebba8-0085-45f5-b3da-fa039307426c","Type":"ContainerDied","Data":"ae96516856a2eea531cf43e68441e43737ec051aa7bf91c34a560c064253d12a"} Feb 27 17:06:05 crc kubenswrapper[4814]: I0227 17:06:05.357057 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536866-qsflg" Feb 27 17:06:05 crc kubenswrapper[4814]: I0227 17:06:05.357069 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae96516856a2eea531cf43e68441e43737ec051aa7bf91c34a560c064253d12a" Feb 27 17:06:05 crc kubenswrapper[4814]: I0227 17:06:05.848066 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536860-n5hhd"] Feb 27 17:06:05 crc kubenswrapper[4814]: I0227 17:06:05.866655 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536860-n5hhd"] Feb 27 17:06:06 crc kubenswrapper[4814]: I0227 17:06:06.504025 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8879d490-f906-4e15-980b-a4bedc4a3196" path="/var/lib/kubelet/pods/8879d490-f906-4e15-980b-a4bedc4a3196/volumes" Feb 27 17:06:09 crc kubenswrapper[4814]: I0227 17:06:09.488367 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:06:09 crc kubenswrapper[4814]: E0227 17:06:09.490416 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:06:23 crc kubenswrapper[4814]: I0227 17:06:23.487571 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:06:23 crc kubenswrapper[4814]: E0227 17:06:23.488343 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:06:34 crc kubenswrapper[4814]: I0227 17:06:34.264979 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-6d4ff56d55-bj24h" podUID="6cc6c3ab-803a-45de-a704-2e180b3bd2ce" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Feb 27 17:06:37 crc kubenswrapper[4814]: I0227 17:06:37.488111 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:06:37 crc kubenswrapper[4814]: E0227 17:06:37.489219 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:06:48 crc kubenswrapper[4814]: I0227 17:06:48.498361 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:06:48 crc kubenswrapper[4814]: E0227 17:06:48.499446 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:06:57 crc kubenswrapper[4814]: I0227 17:06:57.315242 4814 scope.go:117] "RemoveContainer" containerID="a1980c7a0753fa4ef15800e06dd6c6ed370cf7b727c402ba2224547849faac7a" Feb 27 17:07:02 crc kubenswrapper[4814]: I0227 17:07:02.488370 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:07:02 crc kubenswrapper[4814]: E0227 17:07:02.489486 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:07:02 crc kubenswrapper[4814]: I0227 17:07:02.629442 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-77xb5"] Feb 27 17:07:02 crc kubenswrapper[4814]: E0227 17:07:02.630025 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b18ebba8-0085-45f5-b3da-fa039307426c" containerName="oc" Feb 27 17:07:02 crc kubenswrapper[4814]: I0227 17:07:02.630041 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="b18ebba8-0085-45f5-b3da-fa039307426c" containerName="oc" Feb 27 17:07:02 crc kubenswrapper[4814]: I0227 17:07:02.630207 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="b18ebba8-0085-45f5-b3da-fa039307426c" containerName="oc" Feb 27 17:07:02 crc kubenswrapper[4814]: I0227 17:07:02.631525 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:02 crc kubenswrapper[4814]: I0227 17:07:02.667178 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-77xb5"] Feb 27 17:07:02 crc kubenswrapper[4814]: I0227 17:07:02.751311 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-catalog-content\") pod \"redhat-marketplace-77xb5\" (UID: \"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4\") " pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:02 crc kubenswrapper[4814]: I0227 17:07:02.751367 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-utilities\") pod \"redhat-marketplace-77xb5\" (UID: \"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4\") " pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:02 crc kubenswrapper[4814]: I0227 17:07:02.751491 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvcpx\" (UniqueName: \"kubernetes.io/projected/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-kube-api-access-xvcpx\") pod \"redhat-marketplace-77xb5\" (UID: \"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4\") " pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:02 crc kubenswrapper[4814]: I0227 17:07:02.853706 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-catalog-content\") pod \"redhat-marketplace-77xb5\" (UID: \"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4\") " pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:02 crc kubenswrapper[4814]: I0227 17:07:02.853760 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-utilities\") pod \"redhat-marketplace-77xb5\" (UID: \"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4\") " pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:02 crc kubenswrapper[4814]: I0227 17:07:02.853812 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvcpx\" (UniqueName: \"kubernetes.io/projected/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-kube-api-access-xvcpx\") pod \"redhat-marketplace-77xb5\" (UID: \"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4\") " pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:02 crc kubenswrapper[4814]: I0227 17:07:02.854269 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-catalog-content\") pod \"redhat-marketplace-77xb5\" (UID: \"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4\") " pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:02 crc kubenswrapper[4814]: I0227 17:07:02.854604 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-utilities\") pod \"redhat-marketplace-77xb5\" (UID: \"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4\") " pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:02 crc kubenswrapper[4814]: I0227 17:07:02.883137 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvcpx\" (UniqueName: \"kubernetes.io/projected/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-kube-api-access-xvcpx\") pod \"redhat-marketplace-77xb5\" (UID: \"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4\") " pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:02 crc kubenswrapper[4814]: I0227 17:07:02.993635 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:03 crc kubenswrapper[4814]: I0227 17:07:03.032852 4814 generic.go:334] "Generic (PLEG): container finished" podID="2f3a5797-c53e-4195-8e15-f2ba2016c410" containerID="9f0bc6db4ababccc6c692011a3e2a5c3ad1262beea05e57298962c41feabead8" exitCode=0 Feb 27 17:07:03 crc kubenswrapper[4814]: I0227 17:07:03.032901 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" event={"ID":"2f3a5797-c53e-4195-8e15-f2ba2016c410","Type":"ContainerDied","Data":"9f0bc6db4ababccc6c692011a3e2a5c3ad1262beea05e57298962c41feabead8"} Feb 27 17:07:03 crc kubenswrapper[4814]: I0227 17:07:03.486004 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-77xb5"] Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.049236 4814 generic.go:334] "Generic (PLEG): container finished" podID="6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4" containerID="6e8988d799bfbff186582f79b9fcbb401eb2c00175b72e8d64c77180160297d5" exitCode=0 Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.049381 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77xb5" event={"ID":"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4","Type":"ContainerDied","Data":"6e8988d799bfbff186582f79b9fcbb401eb2c00175b72e8d64c77180160297d5"} Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.049802 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77xb5" event={"ID":"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4","Type":"ContainerStarted","Data":"ecf31cd1a7d513a9e9fdf3c5c4f78a5ca40a8bd220385a6567b45e90efe99adc"} Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.564889 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.592111 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-migration-ssh-key-1\") pod \"2f3a5797-c53e-4195-8e15-f2ba2016c410\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.592250 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-combined-ca-bundle\") pod \"2f3a5797-c53e-4195-8e15-f2ba2016c410\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.594889 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-migration-ssh-key-0\") pod \"2f3a5797-c53e-4195-8e15-f2ba2016c410\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.595133 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-inventory\") pod \"2f3a5797-c53e-4195-8e15-f2ba2016c410\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.595159 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-3\") pod \"2f3a5797-c53e-4195-8e15-f2ba2016c410\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.595224 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dd2tj\" (UniqueName: \"kubernetes.io/projected/2f3a5797-c53e-4195-8e15-f2ba2016c410-kube-api-access-dd2tj\") pod \"2f3a5797-c53e-4195-8e15-f2ba2016c410\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.595360 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-2\") pod \"2f3a5797-c53e-4195-8e15-f2ba2016c410\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.595409 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-1\") pod \"2f3a5797-c53e-4195-8e15-f2ba2016c410\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.595457 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-extra-config-0\") pod \"2f3a5797-c53e-4195-8e15-f2ba2016c410\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.595502 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-0\") pod \"2f3a5797-c53e-4195-8e15-f2ba2016c410\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.595523 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-ssh-key-openstack-edpm-ipam\") pod \"2f3a5797-c53e-4195-8e15-f2ba2016c410\" (UID: \"2f3a5797-c53e-4195-8e15-f2ba2016c410\") " Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.597794 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "2f3a5797-c53e-4195-8e15-f2ba2016c410" (UID: "2f3a5797-c53e-4195-8e15-f2ba2016c410"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.615445 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f3a5797-c53e-4195-8e15-f2ba2016c410-kube-api-access-dd2tj" (OuterVolumeSpecName: "kube-api-access-dd2tj") pod "2f3a5797-c53e-4195-8e15-f2ba2016c410" (UID: "2f3a5797-c53e-4195-8e15-f2ba2016c410"). InnerVolumeSpecName "kube-api-access-dd2tj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.637732 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "2f3a5797-c53e-4195-8e15-f2ba2016c410" (UID: "2f3a5797-c53e-4195-8e15-f2ba2016c410"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.644342 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-inventory" (OuterVolumeSpecName: "inventory") pod "2f3a5797-c53e-4195-8e15-f2ba2016c410" (UID: "2f3a5797-c53e-4195-8e15-f2ba2016c410"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.644864 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "2f3a5797-c53e-4195-8e15-f2ba2016c410" (UID: "2f3a5797-c53e-4195-8e15-f2ba2016c410"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.646506 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "2f3a5797-c53e-4195-8e15-f2ba2016c410" (UID: "2f3a5797-c53e-4195-8e15-f2ba2016c410"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.647983 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "2f3a5797-c53e-4195-8e15-f2ba2016c410" (UID: "2f3a5797-c53e-4195-8e15-f2ba2016c410"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.652151 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "2f3a5797-c53e-4195-8e15-f2ba2016c410" (UID: "2f3a5797-c53e-4195-8e15-f2ba2016c410"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.653234 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2f3a5797-c53e-4195-8e15-f2ba2016c410" (UID: "2f3a5797-c53e-4195-8e15-f2ba2016c410"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.656447 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "2f3a5797-c53e-4195-8e15-f2ba2016c410" (UID: "2f3a5797-c53e-4195-8e15-f2ba2016c410"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.664206 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "2f3a5797-c53e-4195-8e15-f2ba2016c410" (UID: "2f3a5797-c53e-4195-8e15-f2ba2016c410"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.698480 4814 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.698538 4814 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.698553 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.698564 4814 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.698575 4814 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.698584 4814 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.698592 4814 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.698602 4814 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.698611 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dd2tj\" (UniqueName: \"kubernetes.io/projected/2f3a5797-c53e-4195-8e15-f2ba2016c410-kube-api-access-dd2tj\") on node \"crc\" DevicePath \"\"" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.698620 4814 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Feb 27 17:07:04 crc kubenswrapper[4814]: I0227 17:07:04.698628 4814 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2f3a5797-c53e-4195-8e15-f2ba2016c410-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.062380 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" event={"ID":"2f3a5797-c53e-4195-8e15-f2ba2016c410","Type":"ContainerDied","Data":"8c8d7c3eda3e5e920f27fc63d41f1e43c8224047bd2abdce7cea2ae46c88e4bd"} Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.062395 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-48n4m" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.062419 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c8d7c3eda3e5e920f27fc63d41f1e43c8224047bd2abdce7cea2ae46c88e4bd" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.069594 4814 generic.go:334] "Generic (PLEG): container finished" podID="6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4" containerID="154d5bb78e351c2242aedce8b5c96d459ff99179c08eb2647e4abf1bba21af9a" exitCode=0 Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.069680 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77xb5" event={"ID":"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4","Type":"ContainerDied","Data":"154d5bb78e351c2242aedce8b5c96d459ff99179c08eb2647e4abf1bba21af9a"} Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.201393 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55"] Feb 27 17:07:05 crc kubenswrapper[4814]: E0227 17:07:05.201840 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3a5797-c53e-4195-8e15-f2ba2016c410" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.201856 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3a5797-c53e-4195-8e15-f2ba2016c410" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.202030 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f3a5797-c53e-4195-8e15-f2ba2016c410" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.202868 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.209188 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.209417 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.209527 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6mrzl" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.209623 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.212918 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.217921 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55"] Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.324410 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh6jj\" (UniqueName: \"kubernetes.io/projected/fee45495-5dbd-4679-b75d-bd5847b4a0fe-kube-api-access-nh6jj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.324509 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.324595 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.324678 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.324721 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.324748 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.324890 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.427132 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.427246 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh6jj\" (UniqueName: \"kubernetes.io/projected/fee45495-5dbd-4679-b75d-bd5847b4a0fe-kube-api-access-nh6jj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.427430 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.427582 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.427652 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.427701 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.427765 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.431357 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.431608 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.432336 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.432379 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.433725 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.433883 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.457145 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh6jj\" (UniqueName: \"kubernetes.io/projected/fee45495-5dbd-4679-b75d-bd5847b4a0fe-kube-api-access-nh6jj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-krg55\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:05 crc kubenswrapper[4814]: I0227 17:07:05.520544 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:07:06 crc kubenswrapper[4814]: I0227 17:07:06.084907 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77xb5" event={"ID":"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4","Type":"ContainerStarted","Data":"03b03acb463d2a74a2bb85f48f5bc654eb041b4aa65a326264e76598993932db"} Feb 27 17:07:06 crc kubenswrapper[4814]: I0227 17:07:06.117217 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-77xb5" podStartSLOduration=2.653161339 podStartE2EDuration="4.117189268s" podCreationTimestamp="2026-02-27 17:07:02 +0000 UTC" firstStartedPulling="2026-02-27 17:07:04.052786131 +0000 UTC m=+2636.505410971" lastFinishedPulling="2026-02-27 17:07:05.51681406 +0000 UTC m=+2637.969438900" observedRunningTime="2026-02-27 17:07:06.106457585 +0000 UTC m=+2638.559082445" watchObservedRunningTime="2026-02-27 17:07:06.117189268 +0000 UTC m=+2638.569814118" Feb 27 17:07:06 crc kubenswrapper[4814]: I0227 17:07:06.150329 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55"] Feb 27 17:07:06 crc kubenswrapper[4814]: W0227 17:07:06.161496 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfee45495_5dbd_4679_b75d_bd5847b4a0fe.slice/crio-cd7a6b0f36722bd977c2d01faa4ef39bd34d1c24ffbec9dd09fffdf38a375578 WatchSource:0}: Error finding container cd7a6b0f36722bd977c2d01faa4ef39bd34d1c24ffbec9dd09fffdf38a375578: Status 404 returned error can't find the container with id cd7a6b0f36722bd977c2d01faa4ef39bd34d1c24ffbec9dd09fffdf38a375578 Feb 27 17:07:07 crc kubenswrapper[4814]: I0227 17:07:07.099553 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" event={"ID":"fee45495-5dbd-4679-b75d-bd5847b4a0fe","Type":"ContainerStarted","Data":"9765cd052ff4a4aad70a5d038d73542f5ff5c7fe9a19c22a3cf720769d26bde2"} Feb 27 17:07:07 crc kubenswrapper[4814]: I0227 17:07:07.100422 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" event={"ID":"fee45495-5dbd-4679-b75d-bd5847b4a0fe","Type":"ContainerStarted","Data":"cd7a6b0f36722bd977c2d01faa4ef39bd34d1c24ffbec9dd09fffdf38a375578"} Feb 27 17:07:07 crc kubenswrapper[4814]: I0227 17:07:07.127831 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" podStartSLOduration=1.6610578230000002 podStartE2EDuration="2.127810392s" podCreationTimestamp="2026-02-27 17:07:05 +0000 UTC" firstStartedPulling="2026-02-27 17:07:06.166661625 +0000 UTC m=+2638.619286495" lastFinishedPulling="2026-02-27 17:07:06.633414244 +0000 UTC m=+2639.086039064" observedRunningTime="2026-02-27 17:07:07.119855005 +0000 UTC m=+2639.572479845" watchObservedRunningTime="2026-02-27 17:07:07.127810392 +0000 UTC m=+2639.580435232" Feb 27 17:07:12 crc kubenswrapper[4814]: I0227 17:07:12.994099 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:12 crc kubenswrapper[4814]: I0227 17:07:12.994896 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:13 crc kubenswrapper[4814]: I0227 17:07:13.069154 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:13 crc kubenswrapper[4814]: I0227 17:07:13.209611 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:13 crc kubenswrapper[4814]: I0227 17:07:13.315075 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-77xb5"] Feb 27 17:07:13 crc kubenswrapper[4814]: I0227 17:07:13.488309 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:07:13 crc kubenswrapper[4814]: E0227 17:07:13.488790 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:07:15 crc kubenswrapper[4814]: I0227 17:07:15.176287 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-77xb5" podUID="6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4" containerName="registry-server" containerID="cri-o://03b03acb463d2a74a2bb85f48f5bc654eb041b4aa65a326264e76598993932db" gracePeriod=2 Feb 27 17:07:15 crc kubenswrapper[4814]: I0227 17:07:15.692390 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:15 crc kubenswrapper[4814]: I0227 17:07:15.858447 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-utilities\") pod \"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4\" (UID: \"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4\") " Feb 27 17:07:15 crc kubenswrapper[4814]: I0227 17:07:15.858611 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvcpx\" (UniqueName: \"kubernetes.io/projected/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-kube-api-access-xvcpx\") pod \"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4\" (UID: \"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4\") " Feb 27 17:07:15 crc kubenswrapper[4814]: I0227 17:07:15.858682 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-catalog-content\") pod \"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4\" (UID: \"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4\") " Feb 27 17:07:15 crc kubenswrapper[4814]: I0227 17:07:15.859876 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-utilities" (OuterVolumeSpecName: "utilities") pod "6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4" (UID: "6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:07:15 crc kubenswrapper[4814]: I0227 17:07:15.868609 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-kube-api-access-xvcpx" (OuterVolumeSpecName: "kube-api-access-xvcpx") pod "6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4" (UID: "6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4"). InnerVolumeSpecName "kube-api-access-xvcpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:07:15 crc kubenswrapper[4814]: I0227 17:07:15.898925 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4" (UID: "6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:07:15 crc kubenswrapper[4814]: I0227 17:07:15.961629 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:07:15 crc kubenswrapper[4814]: I0227 17:07:15.962020 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvcpx\" (UniqueName: \"kubernetes.io/projected/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-kube-api-access-xvcpx\") on node \"crc\" DevicePath \"\"" Feb 27 17:07:15 crc kubenswrapper[4814]: I0227 17:07:15.962042 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:07:16 crc kubenswrapper[4814]: I0227 17:07:16.190901 4814 generic.go:334] "Generic (PLEG): container finished" podID="6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4" containerID="03b03acb463d2a74a2bb85f48f5bc654eb041b4aa65a326264e76598993932db" exitCode=0 Feb 27 17:07:16 crc kubenswrapper[4814]: I0227 17:07:16.190946 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77xb5" event={"ID":"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4","Type":"ContainerDied","Data":"03b03acb463d2a74a2bb85f48f5bc654eb041b4aa65a326264e76598993932db"} Feb 27 17:07:16 crc kubenswrapper[4814]: I0227 17:07:16.190985 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77xb5" event={"ID":"6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4","Type":"ContainerDied","Data":"ecf31cd1a7d513a9e9fdf3c5c4f78a5ca40a8bd220385a6567b45e90efe99adc"} Feb 27 17:07:16 crc kubenswrapper[4814]: I0227 17:07:16.191015 4814 scope.go:117] "RemoveContainer" containerID="03b03acb463d2a74a2bb85f48f5bc654eb041b4aa65a326264e76598993932db" Feb 27 17:07:16 crc kubenswrapper[4814]: I0227 17:07:16.191165 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-77xb5" Feb 27 17:07:16 crc kubenswrapper[4814]: I0227 17:07:16.234077 4814 scope.go:117] "RemoveContainer" containerID="154d5bb78e351c2242aedce8b5c96d459ff99179c08eb2647e4abf1bba21af9a" Feb 27 17:07:16 crc kubenswrapper[4814]: I0227 17:07:16.237098 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-77xb5"] Feb 27 17:07:16 crc kubenswrapper[4814]: I0227 17:07:16.270947 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-77xb5"] Feb 27 17:07:16 crc kubenswrapper[4814]: I0227 17:07:16.304437 4814 scope.go:117] "RemoveContainer" containerID="6e8988d799bfbff186582f79b9fcbb401eb2c00175b72e8d64c77180160297d5" Feb 27 17:07:16 crc kubenswrapper[4814]: I0227 17:07:16.391426 4814 scope.go:117] "RemoveContainer" containerID="03b03acb463d2a74a2bb85f48f5bc654eb041b4aa65a326264e76598993932db" Feb 27 17:07:16 crc kubenswrapper[4814]: E0227 17:07:16.396567 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03b03acb463d2a74a2bb85f48f5bc654eb041b4aa65a326264e76598993932db\": container with ID starting with 03b03acb463d2a74a2bb85f48f5bc654eb041b4aa65a326264e76598993932db not found: ID does not exist" containerID="03b03acb463d2a74a2bb85f48f5bc654eb041b4aa65a326264e76598993932db" Feb 27 17:07:16 crc kubenswrapper[4814]: I0227 17:07:16.396609 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03b03acb463d2a74a2bb85f48f5bc654eb041b4aa65a326264e76598993932db"} err="failed to get container status \"03b03acb463d2a74a2bb85f48f5bc654eb041b4aa65a326264e76598993932db\": rpc error: code = NotFound desc = could not find container \"03b03acb463d2a74a2bb85f48f5bc654eb041b4aa65a326264e76598993932db\": container with ID starting with 03b03acb463d2a74a2bb85f48f5bc654eb041b4aa65a326264e76598993932db not found: ID does not exist" Feb 27 17:07:16 crc kubenswrapper[4814]: I0227 17:07:16.396636 4814 scope.go:117] "RemoveContainer" containerID="154d5bb78e351c2242aedce8b5c96d459ff99179c08eb2647e4abf1bba21af9a" Feb 27 17:07:16 crc kubenswrapper[4814]: E0227 17:07:16.401361 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"154d5bb78e351c2242aedce8b5c96d459ff99179c08eb2647e4abf1bba21af9a\": container with ID starting with 154d5bb78e351c2242aedce8b5c96d459ff99179c08eb2647e4abf1bba21af9a not found: ID does not exist" containerID="154d5bb78e351c2242aedce8b5c96d459ff99179c08eb2647e4abf1bba21af9a" Feb 27 17:07:16 crc kubenswrapper[4814]: I0227 17:07:16.401393 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"154d5bb78e351c2242aedce8b5c96d459ff99179c08eb2647e4abf1bba21af9a"} err="failed to get container status \"154d5bb78e351c2242aedce8b5c96d459ff99179c08eb2647e4abf1bba21af9a\": rpc error: code = NotFound desc = could not find container \"154d5bb78e351c2242aedce8b5c96d459ff99179c08eb2647e4abf1bba21af9a\": container with ID starting with 154d5bb78e351c2242aedce8b5c96d459ff99179c08eb2647e4abf1bba21af9a not found: ID does not exist" Feb 27 17:07:16 crc kubenswrapper[4814]: I0227 17:07:16.401411 4814 scope.go:117] "RemoveContainer" containerID="6e8988d799bfbff186582f79b9fcbb401eb2c00175b72e8d64c77180160297d5" Feb 27 17:07:16 crc kubenswrapper[4814]: E0227 17:07:16.405351 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e8988d799bfbff186582f79b9fcbb401eb2c00175b72e8d64c77180160297d5\": container with ID starting with 6e8988d799bfbff186582f79b9fcbb401eb2c00175b72e8d64c77180160297d5 not found: ID does not exist" containerID="6e8988d799bfbff186582f79b9fcbb401eb2c00175b72e8d64c77180160297d5" Feb 27 17:07:16 crc kubenswrapper[4814]: I0227 17:07:16.405382 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e8988d799bfbff186582f79b9fcbb401eb2c00175b72e8d64c77180160297d5"} err="failed to get container status \"6e8988d799bfbff186582f79b9fcbb401eb2c00175b72e8d64c77180160297d5\": rpc error: code = NotFound desc = could not find container \"6e8988d799bfbff186582f79b9fcbb401eb2c00175b72e8d64c77180160297d5\": container with ID starting with 6e8988d799bfbff186582f79b9fcbb401eb2c00175b72e8d64c77180160297d5 not found: ID does not exist" Feb 27 17:07:16 crc kubenswrapper[4814]: I0227 17:07:16.496766 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4" path="/var/lib/kubelet/pods/6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4/volumes" Feb 27 17:07:27 crc kubenswrapper[4814]: I0227 17:07:27.488356 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:07:27 crc kubenswrapper[4814]: E0227 17:07:27.490762 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:07:42 crc kubenswrapper[4814]: I0227 17:07:42.488381 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:07:42 crc kubenswrapper[4814]: E0227 17:07:42.489749 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:07:56 crc kubenswrapper[4814]: I0227 17:07:56.487772 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:07:56 crc kubenswrapper[4814]: E0227 17:07:56.488792 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:08:00 crc kubenswrapper[4814]: I0227 17:08:00.163657 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536868-8vvj4"] Feb 27 17:08:00 crc kubenswrapper[4814]: E0227 17:08:00.167805 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4" containerName="extract-content" Feb 27 17:08:00 crc kubenswrapper[4814]: I0227 17:08:00.168072 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4" containerName="extract-content" Feb 27 17:08:00 crc kubenswrapper[4814]: E0227 17:08:00.168392 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4" containerName="registry-server" Feb 27 17:08:00 crc kubenswrapper[4814]: I0227 17:08:00.168971 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4" containerName="registry-server" Feb 27 17:08:00 crc kubenswrapper[4814]: E0227 17:08:00.169291 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4" containerName="extract-utilities" Feb 27 17:08:00 crc kubenswrapper[4814]: I0227 17:08:00.169589 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4" containerName="extract-utilities" Feb 27 17:08:00 crc kubenswrapper[4814]: I0227 17:08:00.170310 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fd52f9f-6ca0-4ed0-a455-0ce16452b4c4" containerName="registry-server" Feb 27 17:08:00 crc kubenswrapper[4814]: I0227 17:08:00.171551 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536868-8vvj4" Feb 27 17:08:00 crc kubenswrapper[4814]: I0227 17:08:00.172498 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536868-8vvj4"] Feb 27 17:08:00 crc kubenswrapper[4814]: I0227 17:08:00.175415 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:08:00 crc kubenswrapper[4814]: I0227 17:08:00.175556 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:08:00 crc kubenswrapper[4814]: I0227 17:08:00.175924 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:08:00 crc kubenswrapper[4814]: I0227 17:08:00.285076 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phsjs\" (UniqueName: \"kubernetes.io/projected/fb327adc-9d58-4183-9e6e-866708bbae18-kube-api-access-phsjs\") pod \"auto-csr-approver-29536868-8vvj4\" (UID: \"fb327adc-9d58-4183-9e6e-866708bbae18\") " pod="openshift-infra/auto-csr-approver-29536868-8vvj4" Feb 27 17:08:00 crc kubenswrapper[4814]: I0227 17:08:00.387073 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phsjs\" (UniqueName: \"kubernetes.io/projected/fb327adc-9d58-4183-9e6e-866708bbae18-kube-api-access-phsjs\") pod \"auto-csr-approver-29536868-8vvj4\" (UID: \"fb327adc-9d58-4183-9e6e-866708bbae18\") " pod="openshift-infra/auto-csr-approver-29536868-8vvj4" Feb 27 17:08:00 crc kubenswrapper[4814]: I0227 17:08:00.423280 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phsjs\" (UniqueName: \"kubernetes.io/projected/fb327adc-9d58-4183-9e6e-866708bbae18-kube-api-access-phsjs\") pod \"auto-csr-approver-29536868-8vvj4\" (UID: \"fb327adc-9d58-4183-9e6e-866708bbae18\") " pod="openshift-infra/auto-csr-approver-29536868-8vvj4" Feb 27 17:08:00 crc kubenswrapper[4814]: I0227 17:08:00.508458 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536868-8vvj4" Feb 27 17:08:01 crc kubenswrapper[4814]: I0227 17:08:01.042204 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536868-8vvj4"] Feb 27 17:08:01 crc kubenswrapper[4814]: I0227 17:08:01.710527 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536868-8vvj4" event={"ID":"fb327adc-9d58-4183-9e6e-866708bbae18","Type":"ContainerStarted","Data":"1b3f5cec2d69b54fc92225f0eee8732f0e5fc29d97dd6ee3b9e7ab762f22a1b4"} Feb 27 17:08:02 crc kubenswrapper[4814]: I0227 17:08:02.721647 4814 generic.go:334] "Generic (PLEG): container finished" podID="fb327adc-9d58-4183-9e6e-866708bbae18" containerID="873c09fc96c2432ed431fedafa4ebb9be4dcc6d0b95ee8eb9f59810ee5f03253" exitCode=0 Feb 27 17:08:02 crc kubenswrapper[4814]: I0227 17:08:02.721785 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536868-8vvj4" event={"ID":"fb327adc-9d58-4183-9e6e-866708bbae18","Type":"ContainerDied","Data":"873c09fc96c2432ed431fedafa4ebb9be4dcc6d0b95ee8eb9f59810ee5f03253"} Feb 27 17:08:04 crc kubenswrapper[4814]: I0227 17:08:04.136869 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536868-8vvj4" Feb 27 17:08:04 crc kubenswrapper[4814]: I0227 17:08:04.301759 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phsjs\" (UniqueName: \"kubernetes.io/projected/fb327adc-9d58-4183-9e6e-866708bbae18-kube-api-access-phsjs\") pod \"fb327adc-9d58-4183-9e6e-866708bbae18\" (UID: \"fb327adc-9d58-4183-9e6e-866708bbae18\") " Feb 27 17:08:04 crc kubenswrapper[4814]: I0227 17:08:04.310821 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb327adc-9d58-4183-9e6e-866708bbae18-kube-api-access-phsjs" (OuterVolumeSpecName: "kube-api-access-phsjs") pod "fb327adc-9d58-4183-9e6e-866708bbae18" (UID: "fb327adc-9d58-4183-9e6e-866708bbae18"). InnerVolumeSpecName "kube-api-access-phsjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:08:04 crc kubenswrapper[4814]: I0227 17:08:04.404490 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phsjs\" (UniqueName: \"kubernetes.io/projected/fb327adc-9d58-4183-9e6e-866708bbae18-kube-api-access-phsjs\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:04 crc kubenswrapper[4814]: I0227 17:08:04.747231 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536868-8vvj4" event={"ID":"fb327adc-9d58-4183-9e6e-866708bbae18","Type":"ContainerDied","Data":"1b3f5cec2d69b54fc92225f0eee8732f0e5fc29d97dd6ee3b9e7ab762f22a1b4"} Feb 27 17:08:04 crc kubenswrapper[4814]: I0227 17:08:04.747321 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b3f5cec2d69b54fc92225f0eee8732f0e5fc29d97dd6ee3b9e7ab762f22a1b4" Feb 27 17:08:04 crc kubenswrapper[4814]: I0227 17:08:04.747378 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536868-8vvj4" Feb 27 17:08:05 crc kubenswrapper[4814]: I0227 17:08:05.245779 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536862-b2l7p"] Feb 27 17:08:05 crc kubenswrapper[4814]: I0227 17:08:05.255095 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536862-b2l7p"] Feb 27 17:08:06 crc kubenswrapper[4814]: I0227 17:08:06.499656 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04d54354-19e5-4d8e-9899-eeb32b752e05" path="/var/lib/kubelet/pods/04d54354-19e5-4d8e-9899-eeb32b752e05/volumes" Feb 27 17:08:10 crc kubenswrapper[4814]: I0227 17:08:10.487608 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:08:10 crc kubenswrapper[4814]: E0227 17:08:10.488498 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:08:21 crc kubenswrapper[4814]: I0227 17:08:21.488694 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:08:21 crc kubenswrapper[4814]: E0227 17:08:21.489854 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:08:33 crc kubenswrapper[4814]: I0227 17:08:33.487855 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:08:33 crc kubenswrapper[4814]: E0227 17:08:33.489082 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:08:47 crc kubenswrapper[4814]: I0227 17:08:47.488156 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:08:47 crc kubenswrapper[4814]: E0227 17:08:47.491244 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:08:57 crc kubenswrapper[4814]: I0227 17:08:57.454301 4814 scope.go:117] "RemoveContainer" containerID="0466b5febadec44c8ae866fef3e97d7cd950a179d4ff5a4bdebbf51f9b4a7970" Feb 27 17:08:59 crc kubenswrapper[4814]: I0227 17:08:59.488146 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:08:59 crc kubenswrapper[4814]: E0227 17:08:59.488624 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:09:10 crc kubenswrapper[4814]: I0227 17:09:10.488296 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:09:10 crc kubenswrapper[4814]: E0227 17:09:10.489152 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:09:22 crc kubenswrapper[4814]: I0227 17:09:22.488224 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:09:22 crc kubenswrapper[4814]: E0227 17:09:22.489426 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:09:35 crc kubenswrapper[4814]: I0227 17:09:35.487840 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:09:35 crc kubenswrapper[4814]: E0227 17:09:35.489015 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:09:44 crc kubenswrapper[4814]: I0227 17:09:44.893075 4814 generic.go:334] "Generic (PLEG): container finished" podID="fee45495-5dbd-4679-b75d-bd5847b4a0fe" containerID="9765cd052ff4a4aad70a5d038d73542f5ff5c7fe9a19c22a3cf720769d26bde2" exitCode=0 Feb 27 17:09:44 crc kubenswrapper[4814]: I0227 17:09:44.893190 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" event={"ID":"fee45495-5dbd-4679-b75d-bd5847b4a0fe","Type":"ContainerDied","Data":"9765cd052ff4a4aad70a5d038d73542f5ff5c7fe9a19c22a3cf720769d26bde2"} Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.446715 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.575686 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-inventory\") pod \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.575784 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ssh-key-openstack-edpm-ipam\") pod \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.575817 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-2\") pod \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.575872 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-0\") pod \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.575971 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nh6jj\" (UniqueName: \"kubernetes.io/projected/fee45495-5dbd-4679-b75d-bd5847b4a0fe-kube-api-access-nh6jj\") pod \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.576103 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-1\") pod \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.576136 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-telemetry-combined-ca-bundle\") pod \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\" (UID: \"fee45495-5dbd-4679-b75d-bd5847b4a0fe\") " Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.582862 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "fee45495-5dbd-4679-b75d-bd5847b4a0fe" (UID: "fee45495-5dbd-4679-b75d-bd5847b4a0fe"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.583474 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fee45495-5dbd-4679-b75d-bd5847b4a0fe-kube-api-access-nh6jj" (OuterVolumeSpecName: "kube-api-access-nh6jj") pod "fee45495-5dbd-4679-b75d-bd5847b4a0fe" (UID: "fee45495-5dbd-4679-b75d-bd5847b4a0fe"). InnerVolumeSpecName "kube-api-access-nh6jj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.607726 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "fee45495-5dbd-4679-b75d-bd5847b4a0fe" (UID: "fee45495-5dbd-4679-b75d-bd5847b4a0fe"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.617469 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "fee45495-5dbd-4679-b75d-bd5847b4a0fe" (UID: "fee45495-5dbd-4679-b75d-bd5847b4a0fe"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.628377 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "fee45495-5dbd-4679-b75d-bd5847b4a0fe" (UID: "fee45495-5dbd-4679-b75d-bd5847b4a0fe"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.641239 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-inventory" (OuterVolumeSpecName: "inventory") pod "fee45495-5dbd-4679-b75d-bd5847b4a0fe" (UID: "fee45495-5dbd-4679-b75d-bd5847b4a0fe"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.642186 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "fee45495-5dbd-4679-b75d-bd5847b4a0fe" (UID: "fee45495-5dbd-4679-b75d-bd5847b4a0fe"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.679743 4814 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.680749 4814 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.680809 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.680832 4814 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.680853 4814 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.680874 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nh6jj\" (UniqueName: \"kubernetes.io/projected/fee45495-5dbd-4679-b75d-bd5847b4a0fe-kube-api-access-nh6jj\") on node \"crc\" DevicePath \"\"" Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.680895 4814 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fee45495-5dbd-4679-b75d-bd5847b4a0fe-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.914968 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" event={"ID":"fee45495-5dbd-4679-b75d-bd5847b4a0fe","Type":"ContainerDied","Data":"cd7a6b0f36722bd977c2d01faa4ef39bd34d1c24ffbec9dd09fffdf38a375578"} Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.915299 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd7a6b0f36722bd977c2d01faa4ef39bd34d1c24ffbec9dd09fffdf38a375578" Feb 27 17:09:46 crc kubenswrapper[4814]: I0227 17:09:46.915169 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-krg55" Feb 27 17:09:48 crc kubenswrapper[4814]: I0227 17:09:48.493811 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:09:48 crc kubenswrapper[4814]: E0227 17:09:48.494357 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:09:52 crc kubenswrapper[4814]: I0227 17:09:52.989834 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qwzlw"] Feb 27 17:09:52 crc kubenswrapper[4814]: E0227 17:09:52.990602 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb327adc-9d58-4183-9e6e-866708bbae18" containerName="oc" Feb 27 17:09:52 crc kubenswrapper[4814]: I0227 17:09:52.990615 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb327adc-9d58-4183-9e6e-866708bbae18" containerName="oc" Feb 27 17:09:52 crc kubenswrapper[4814]: E0227 17:09:52.990628 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fee45495-5dbd-4679-b75d-bd5847b4a0fe" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 27 17:09:52 crc kubenswrapper[4814]: I0227 17:09:52.990634 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="fee45495-5dbd-4679-b75d-bd5847b4a0fe" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 27 17:09:52 crc kubenswrapper[4814]: I0227 17:09:52.990813 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="fee45495-5dbd-4679-b75d-bd5847b4a0fe" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 27 17:09:52 crc kubenswrapper[4814]: I0227 17:09:52.990827 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb327adc-9d58-4183-9e6e-866708bbae18" containerName="oc" Feb 27 17:09:52 crc kubenswrapper[4814]: I0227 17:09:52.992306 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:09:53 crc kubenswrapper[4814]: I0227 17:09:53.017679 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qwzlw"] Feb 27 17:09:53 crc kubenswrapper[4814]: I0227 17:09:53.023977 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvcpl\" (UniqueName: \"kubernetes.io/projected/92ee499b-b40e-4367-b7df-cf968ff8bf56-kube-api-access-rvcpl\") pod \"redhat-operators-qwzlw\" (UID: \"92ee499b-b40e-4367-b7df-cf968ff8bf56\") " pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:09:53 crc kubenswrapper[4814]: I0227 17:09:53.024022 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ee499b-b40e-4367-b7df-cf968ff8bf56-utilities\") pod \"redhat-operators-qwzlw\" (UID: \"92ee499b-b40e-4367-b7df-cf968ff8bf56\") " pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:09:53 crc kubenswrapper[4814]: I0227 17:09:53.024079 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ee499b-b40e-4367-b7df-cf968ff8bf56-catalog-content\") pod \"redhat-operators-qwzlw\" (UID: \"92ee499b-b40e-4367-b7df-cf968ff8bf56\") " pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:09:53 crc kubenswrapper[4814]: I0227 17:09:53.125917 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvcpl\" (UniqueName: \"kubernetes.io/projected/92ee499b-b40e-4367-b7df-cf968ff8bf56-kube-api-access-rvcpl\") pod \"redhat-operators-qwzlw\" (UID: \"92ee499b-b40e-4367-b7df-cf968ff8bf56\") " pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:09:53 crc kubenswrapper[4814]: I0227 17:09:53.125978 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ee499b-b40e-4367-b7df-cf968ff8bf56-utilities\") pod \"redhat-operators-qwzlw\" (UID: \"92ee499b-b40e-4367-b7df-cf968ff8bf56\") " pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:09:53 crc kubenswrapper[4814]: I0227 17:09:53.126009 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ee499b-b40e-4367-b7df-cf968ff8bf56-catalog-content\") pod \"redhat-operators-qwzlw\" (UID: \"92ee499b-b40e-4367-b7df-cf968ff8bf56\") " pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:09:53 crc kubenswrapper[4814]: I0227 17:09:53.126523 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ee499b-b40e-4367-b7df-cf968ff8bf56-utilities\") pod \"redhat-operators-qwzlw\" (UID: \"92ee499b-b40e-4367-b7df-cf968ff8bf56\") " pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:09:53 crc kubenswrapper[4814]: I0227 17:09:53.126654 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ee499b-b40e-4367-b7df-cf968ff8bf56-catalog-content\") pod \"redhat-operators-qwzlw\" (UID: \"92ee499b-b40e-4367-b7df-cf968ff8bf56\") " pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:09:53 crc kubenswrapper[4814]: I0227 17:09:53.152920 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvcpl\" (UniqueName: \"kubernetes.io/projected/92ee499b-b40e-4367-b7df-cf968ff8bf56-kube-api-access-rvcpl\") pod \"redhat-operators-qwzlw\" (UID: \"92ee499b-b40e-4367-b7df-cf968ff8bf56\") " pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:09:53 crc kubenswrapper[4814]: I0227 17:09:53.321491 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:09:53 crc kubenswrapper[4814]: I0227 17:09:53.793381 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qwzlw"] Feb 27 17:09:53 crc kubenswrapper[4814]: I0227 17:09:53.993576 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwzlw" event={"ID":"92ee499b-b40e-4367-b7df-cf968ff8bf56","Type":"ContainerStarted","Data":"ac470514637d9e862d15f00a2da99ccc77a93563ec3a1e84a66ac0fe24868902"} Feb 27 17:09:53 crc kubenswrapper[4814]: I0227 17:09:53.993818 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwzlw" event={"ID":"92ee499b-b40e-4367-b7df-cf968ff8bf56","Type":"ContainerStarted","Data":"61655c641cbfc9a7e31ad24f54a68825a1f2ad277c62087ef00f8e5d6147606e"} Feb 27 17:09:55 crc kubenswrapper[4814]: I0227 17:09:55.012934 4814 generic.go:334] "Generic (PLEG): container finished" podID="92ee499b-b40e-4367-b7df-cf968ff8bf56" containerID="ac470514637d9e862d15f00a2da99ccc77a93563ec3a1e84a66ac0fe24868902" exitCode=0 Feb 27 17:09:55 crc kubenswrapper[4814]: I0227 17:09:55.013058 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwzlw" event={"ID":"92ee499b-b40e-4367-b7df-cf968ff8bf56","Type":"ContainerDied","Data":"ac470514637d9e862d15f00a2da99ccc77a93563ec3a1e84a66ac0fe24868902"} Feb 27 17:09:55 crc kubenswrapper[4814]: I0227 17:09:55.018605 4814 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:09:56 crc kubenswrapper[4814]: I0227 17:09:56.024423 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwzlw" event={"ID":"92ee499b-b40e-4367-b7df-cf968ff8bf56","Type":"ContainerStarted","Data":"4367e73b53890053072dcb9cce5d78215fb8085029dfd8faa226de6b916a50e4"} Feb 27 17:09:58 crc kubenswrapper[4814]: I0227 17:09:58.047178 4814 generic.go:334] "Generic (PLEG): container finished" podID="92ee499b-b40e-4367-b7df-cf968ff8bf56" containerID="4367e73b53890053072dcb9cce5d78215fb8085029dfd8faa226de6b916a50e4" exitCode=0 Feb 27 17:09:58 crc kubenswrapper[4814]: I0227 17:09:58.047307 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwzlw" event={"ID":"92ee499b-b40e-4367-b7df-cf968ff8bf56","Type":"ContainerDied","Data":"4367e73b53890053072dcb9cce5d78215fb8085029dfd8faa226de6b916a50e4"} Feb 27 17:10:00 crc kubenswrapper[4814]: I0227 17:10:00.083989 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwzlw" event={"ID":"92ee499b-b40e-4367-b7df-cf968ff8bf56","Type":"ContainerStarted","Data":"5ff9540cb6008b5c63b94b66d6c8434b4fa790299ee6dc182131521d3baaf1df"} Feb 27 17:10:00 crc kubenswrapper[4814]: I0227 17:10:00.122353 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qwzlw" podStartSLOduration=4.363372433 podStartE2EDuration="8.122330619s" podCreationTimestamp="2026-02-27 17:09:52 +0000 UTC" firstStartedPulling="2026-02-27 17:09:55.018016188 +0000 UTC m=+2807.470641058" lastFinishedPulling="2026-02-27 17:09:58.776974404 +0000 UTC m=+2811.229599244" observedRunningTime="2026-02-27 17:10:00.116176928 +0000 UTC m=+2812.568801768" watchObservedRunningTime="2026-02-27 17:10:00.122330619 +0000 UTC m=+2812.574955459" Feb 27 17:10:00 crc kubenswrapper[4814]: I0227 17:10:00.174910 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536870-wdkzj"] Feb 27 17:10:00 crc kubenswrapper[4814]: I0227 17:10:00.178152 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536870-wdkzj" Feb 27 17:10:00 crc kubenswrapper[4814]: I0227 17:10:00.182432 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:10:00 crc kubenswrapper[4814]: I0227 17:10:00.188049 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:10:00 crc kubenswrapper[4814]: I0227 17:10:00.188409 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:10:00 crc kubenswrapper[4814]: I0227 17:10:00.188966 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536870-wdkzj"] Feb 27 17:10:00 crc kubenswrapper[4814]: I0227 17:10:00.286344 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6txnx\" (UniqueName: \"kubernetes.io/projected/33af9720-6802-4c2b-8fbf-37ec02ece573-kube-api-access-6txnx\") pod \"auto-csr-approver-29536870-wdkzj\" (UID: \"33af9720-6802-4c2b-8fbf-37ec02ece573\") " pod="openshift-infra/auto-csr-approver-29536870-wdkzj" Feb 27 17:10:00 crc kubenswrapper[4814]: I0227 17:10:00.387935 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6txnx\" (UniqueName: \"kubernetes.io/projected/33af9720-6802-4c2b-8fbf-37ec02ece573-kube-api-access-6txnx\") pod \"auto-csr-approver-29536870-wdkzj\" (UID: \"33af9720-6802-4c2b-8fbf-37ec02ece573\") " pod="openshift-infra/auto-csr-approver-29536870-wdkzj" Feb 27 17:10:00 crc kubenswrapper[4814]: I0227 17:10:00.408557 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6txnx\" (UniqueName: \"kubernetes.io/projected/33af9720-6802-4c2b-8fbf-37ec02ece573-kube-api-access-6txnx\") pod \"auto-csr-approver-29536870-wdkzj\" (UID: \"33af9720-6802-4c2b-8fbf-37ec02ece573\") " pod="openshift-infra/auto-csr-approver-29536870-wdkzj" Feb 27 17:10:00 crc kubenswrapper[4814]: I0227 17:10:00.487879 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:10:00 crc kubenswrapper[4814]: I0227 17:10:00.510640 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536870-wdkzj" Feb 27 17:10:01 crc kubenswrapper[4814]: I0227 17:10:01.016109 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536870-wdkzj"] Feb 27 17:10:01 crc kubenswrapper[4814]: I0227 17:10:01.096245 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536870-wdkzj" event={"ID":"33af9720-6802-4c2b-8fbf-37ec02ece573","Type":"ContainerStarted","Data":"579a7e38b056cba401001ed8118aa95ddb11db7f5e490ac404f20624c6e042c0"} Feb 27 17:10:01 crc kubenswrapper[4814]: I0227 17:10:01.099572 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"13fbd8981cbf65fbf38caf5962c43492d7821a7f2bb43648de09196463587955"} Feb 27 17:10:03 crc kubenswrapper[4814]: I0227 17:10:03.119567 4814 generic.go:334] "Generic (PLEG): container finished" podID="33af9720-6802-4c2b-8fbf-37ec02ece573" containerID="e74711f48be7b2944ceec9791efefe4bbba67956665a3f3bc133bd0e81e84260" exitCode=0 Feb 27 17:10:03 crc kubenswrapper[4814]: I0227 17:10:03.119669 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536870-wdkzj" event={"ID":"33af9720-6802-4c2b-8fbf-37ec02ece573","Type":"ContainerDied","Data":"e74711f48be7b2944ceec9791efefe4bbba67956665a3f3bc133bd0e81e84260"} Feb 27 17:10:03 crc kubenswrapper[4814]: I0227 17:10:03.322138 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:10:03 crc kubenswrapper[4814]: I0227 17:10:03.322609 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:10:04 crc kubenswrapper[4814]: I0227 17:10:04.390670 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qwzlw" podUID="92ee499b-b40e-4367-b7df-cf968ff8bf56" containerName="registry-server" probeResult="failure" output=< Feb 27 17:10:04 crc kubenswrapper[4814]: timeout: failed to connect service ":50051" within 1s Feb 27 17:10:04 crc kubenswrapper[4814]: > Feb 27 17:10:04 crc kubenswrapper[4814]: I0227 17:10:04.527070 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536870-wdkzj" Feb 27 17:10:04 crc kubenswrapper[4814]: I0227 17:10:04.591705 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6txnx\" (UniqueName: \"kubernetes.io/projected/33af9720-6802-4c2b-8fbf-37ec02ece573-kube-api-access-6txnx\") pod \"33af9720-6802-4c2b-8fbf-37ec02ece573\" (UID: \"33af9720-6802-4c2b-8fbf-37ec02ece573\") " Feb 27 17:10:04 crc kubenswrapper[4814]: I0227 17:10:04.597651 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33af9720-6802-4c2b-8fbf-37ec02ece573-kube-api-access-6txnx" (OuterVolumeSpecName: "kube-api-access-6txnx") pod "33af9720-6802-4c2b-8fbf-37ec02ece573" (UID: "33af9720-6802-4c2b-8fbf-37ec02ece573"). InnerVolumeSpecName "kube-api-access-6txnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:10:04 crc kubenswrapper[4814]: I0227 17:10:04.694335 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6txnx\" (UniqueName: \"kubernetes.io/projected/33af9720-6802-4c2b-8fbf-37ec02ece573-kube-api-access-6txnx\") on node \"crc\" DevicePath \"\"" Feb 27 17:10:05 crc kubenswrapper[4814]: I0227 17:10:05.143388 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536870-wdkzj" event={"ID":"33af9720-6802-4c2b-8fbf-37ec02ece573","Type":"ContainerDied","Data":"579a7e38b056cba401001ed8118aa95ddb11db7f5e490ac404f20624c6e042c0"} Feb 27 17:10:05 crc kubenswrapper[4814]: I0227 17:10:05.143673 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="579a7e38b056cba401001ed8118aa95ddb11db7f5e490ac404f20624c6e042c0" Feb 27 17:10:05 crc kubenswrapper[4814]: I0227 17:10:05.143480 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536870-wdkzj" Feb 27 17:10:05 crc kubenswrapper[4814]: I0227 17:10:05.617774 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536864-w6vcs"] Feb 27 17:10:05 crc kubenswrapper[4814]: I0227 17:10:05.630647 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536864-w6vcs"] Feb 27 17:10:06 crc kubenswrapper[4814]: I0227 17:10:06.507013 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19875f7e-5900-4480-af38-cb7e970b67d2" path="/var/lib/kubelet/pods/19875f7e-5900-4480-af38-cb7e970b67d2/volumes" Feb 27 17:10:13 crc kubenswrapper[4814]: I0227 17:10:13.393869 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:10:13 crc kubenswrapper[4814]: I0227 17:10:13.460443 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:10:16 crc kubenswrapper[4814]: I0227 17:10:16.095126 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qwzlw"] Feb 27 17:10:16 crc kubenswrapper[4814]: I0227 17:10:16.095849 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qwzlw" podUID="92ee499b-b40e-4367-b7df-cf968ff8bf56" containerName="registry-server" containerID="cri-o://5ff9540cb6008b5c63b94b66d6c8434b4fa790299ee6dc182131521d3baaf1df" gracePeriod=2 Feb 27 17:10:16 crc kubenswrapper[4814]: I0227 17:10:16.278171 4814 generic.go:334] "Generic (PLEG): container finished" podID="92ee499b-b40e-4367-b7df-cf968ff8bf56" containerID="5ff9540cb6008b5c63b94b66d6c8434b4fa790299ee6dc182131521d3baaf1df" exitCode=0 Feb 27 17:10:16 crc kubenswrapper[4814]: I0227 17:10:16.278318 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwzlw" event={"ID":"92ee499b-b40e-4367-b7df-cf968ff8bf56","Type":"ContainerDied","Data":"5ff9540cb6008b5c63b94b66d6c8434b4fa790299ee6dc182131521d3baaf1df"} Feb 27 17:10:16 crc kubenswrapper[4814]: I0227 17:10:16.646997 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:10:16 crc kubenswrapper[4814]: I0227 17:10:16.772059 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ee499b-b40e-4367-b7df-cf968ff8bf56-catalog-content\") pod \"92ee499b-b40e-4367-b7df-cf968ff8bf56\" (UID: \"92ee499b-b40e-4367-b7df-cf968ff8bf56\") " Feb 27 17:10:16 crc kubenswrapper[4814]: I0227 17:10:16.772755 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ee499b-b40e-4367-b7df-cf968ff8bf56-utilities\") pod \"92ee499b-b40e-4367-b7df-cf968ff8bf56\" (UID: \"92ee499b-b40e-4367-b7df-cf968ff8bf56\") " Feb 27 17:10:16 crc kubenswrapper[4814]: I0227 17:10:16.772857 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvcpl\" (UniqueName: \"kubernetes.io/projected/92ee499b-b40e-4367-b7df-cf968ff8bf56-kube-api-access-rvcpl\") pod \"92ee499b-b40e-4367-b7df-cf968ff8bf56\" (UID: \"92ee499b-b40e-4367-b7df-cf968ff8bf56\") " Feb 27 17:10:16 crc kubenswrapper[4814]: I0227 17:10:16.773867 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92ee499b-b40e-4367-b7df-cf968ff8bf56-utilities" (OuterVolumeSpecName: "utilities") pod "92ee499b-b40e-4367-b7df-cf968ff8bf56" (UID: "92ee499b-b40e-4367-b7df-cf968ff8bf56"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:10:16 crc kubenswrapper[4814]: I0227 17:10:16.791543 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92ee499b-b40e-4367-b7df-cf968ff8bf56-kube-api-access-rvcpl" (OuterVolumeSpecName: "kube-api-access-rvcpl") pod "92ee499b-b40e-4367-b7df-cf968ff8bf56" (UID: "92ee499b-b40e-4367-b7df-cf968ff8bf56"). InnerVolumeSpecName "kube-api-access-rvcpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:10:16 crc kubenswrapper[4814]: I0227 17:10:16.876779 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ee499b-b40e-4367-b7df-cf968ff8bf56-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:10:16 crc kubenswrapper[4814]: I0227 17:10:16.876821 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvcpl\" (UniqueName: \"kubernetes.io/projected/92ee499b-b40e-4367-b7df-cf968ff8bf56-kube-api-access-rvcpl\") on node \"crc\" DevicePath \"\"" Feb 27 17:10:16 crc kubenswrapper[4814]: I0227 17:10:16.948776 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92ee499b-b40e-4367-b7df-cf968ff8bf56-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92ee499b-b40e-4367-b7df-cf968ff8bf56" (UID: "92ee499b-b40e-4367-b7df-cf968ff8bf56"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:10:16 crc kubenswrapper[4814]: I0227 17:10:16.979008 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ee499b-b40e-4367-b7df-cf968ff8bf56-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:10:17 crc kubenswrapper[4814]: I0227 17:10:17.299167 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwzlw" event={"ID":"92ee499b-b40e-4367-b7df-cf968ff8bf56","Type":"ContainerDied","Data":"61655c641cbfc9a7e31ad24f54a68825a1f2ad277c62087ef00f8e5d6147606e"} Feb 27 17:10:17 crc kubenswrapper[4814]: I0227 17:10:17.299248 4814 scope.go:117] "RemoveContainer" containerID="5ff9540cb6008b5c63b94b66d6c8434b4fa790299ee6dc182131521d3baaf1df" Feb 27 17:10:17 crc kubenswrapper[4814]: I0227 17:10:17.300461 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwzlw" Feb 27 17:10:17 crc kubenswrapper[4814]: I0227 17:10:17.324850 4814 scope.go:117] "RemoveContainer" containerID="4367e73b53890053072dcb9cce5d78215fb8085029dfd8faa226de6b916a50e4" Feb 27 17:10:17 crc kubenswrapper[4814]: I0227 17:10:17.359470 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qwzlw"] Feb 27 17:10:17 crc kubenswrapper[4814]: I0227 17:10:17.367749 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qwzlw"] Feb 27 17:10:17 crc kubenswrapper[4814]: I0227 17:10:17.387155 4814 scope.go:117] "RemoveContainer" containerID="ac470514637d9e862d15f00a2da99ccc77a93563ec3a1e84a66ac0fe24868902" Feb 27 17:10:18 crc kubenswrapper[4814]: I0227 17:10:18.505201 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92ee499b-b40e-4367-b7df-cf968ff8bf56" path="/var/lib/kubelet/pods/92ee499b-b40e-4367-b7df-cf968ff8bf56/volumes" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.085629 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Feb 27 17:10:30 crc kubenswrapper[4814]: E0227 17:10:30.087206 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92ee499b-b40e-4367-b7df-cf968ff8bf56" containerName="registry-server" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.087234 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="92ee499b-b40e-4367-b7df-cf968ff8bf56" containerName="registry-server" Feb 27 17:10:30 crc kubenswrapper[4814]: E0227 17:10:30.087285 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33af9720-6802-4c2b-8fbf-37ec02ece573" containerName="oc" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.087298 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="33af9720-6802-4c2b-8fbf-37ec02ece573" containerName="oc" Feb 27 17:10:30 crc kubenswrapper[4814]: E0227 17:10:30.087342 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92ee499b-b40e-4367-b7df-cf968ff8bf56" containerName="extract-content" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.087355 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="92ee499b-b40e-4367-b7df-cf968ff8bf56" containerName="extract-content" Feb 27 17:10:30 crc kubenswrapper[4814]: E0227 17:10:30.087408 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92ee499b-b40e-4367-b7df-cf968ff8bf56" containerName="extract-utilities" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.087421 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="92ee499b-b40e-4367-b7df-cf968ff8bf56" containerName="extract-utilities" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.087782 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="92ee499b-b40e-4367-b7df-cf968ff8bf56" containerName="registry-server" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.087818 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="33af9720-6802-4c2b-8fbf-37ec02ece573" containerName="oc" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.088940 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.094537 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.094795 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cvqcf" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.095224 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.095486 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.129737 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.170038 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c11159aa-3255-4574-81a3-182c8a30d970-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.170099 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c11159aa-3255-4574-81a3-182c8a30d970-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.170613 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.170751 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.170870 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c11159aa-3255-4574-81a3-182c8a30d970-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.170902 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.171065 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c11159aa-3255-4574-81a3-182c8a30d970-config-data\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.171137 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r85q\" (UniqueName: \"kubernetes.io/projected/c11159aa-3255-4574-81a3-182c8a30d970-kube-api-access-9r85q\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.171215 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.275829 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r85q\" (UniqueName: \"kubernetes.io/projected/c11159aa-3255-4574-81a3-182c8a30d970-kube-api-access-9r85q\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.275924 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.276067 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c11159aa-3255-4574-81a3-182c8a30d970-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.276124 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c11159aa-3255-4574-81a3-182c8a30d970-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.276234 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.276308 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.276366 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c11159aa-3255-4574-81a3-182c8a30d970-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.276400 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.276474 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c11159aa-3255-4574-81a3-182c8a30d970-config-data\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.276689 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c11159aa-3255-4574-81a3-182c8a30d970-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.277016 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c11159aa-3255-4574-81a3-182c8a30d970-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.277779 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.278503 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c11159aa-3255-4574-81a3-182c8a30d970-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.278726 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c11159aa-3255-4574-81a3-182c8a30d970-config-data\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.282860 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.289165 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.290455 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.295602 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r85q\" (UniqueName: \"kubernetes.io/projected/c11159aa-3255-4574-81a3-182c8a30d970-kube-api-access-9r85q\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.317459 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.431152 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 27 17:10:30 crc kubenswrapper[4814]: I0227 17:10:30.870874 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 27 17:10:30 crc kubenswrapper[4814]: W0227 17:10:30.887808 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc11159aa_3255_4574_81a3_182c8a30d970.slice/crio-fb542b3688380e6d3c1742f0ae5c7794c81dc270fe03e7866576ac197c341bd9 WatchSource:0}: Error finding container fb542b3688380e6d3c1742f0ae5c7794c81dc270fe03e7866576ac197c341bd9: Status 404 returned error can't find the container with id fb542b3688380e6d3c1742f0ae5c7794c81dc270fe03e7866576ac197c341bd9 Feb 27 17:10:31 crc kubenswrapper[4814]: I0227 17:10:31.468278 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c11159aa-3255-4574-81a3-182c8a30d970","Type":"ContainerStarted","Data":"fb542b3688380e6d3c1742f0ae5c7794c81dc270fe03e7866576ac197c341bd9"} Feb 27 17:10:57 crc kubenswrapper[4814]: I0227 17:10:57.576405 4814 scope.go:117] "RemoveContainer" containerID="9b7af82cc2107f82111d248255ef0d73a7e8d5ec24fb6ce2e86535ed73642963" Feb 27 17:11:06 crc kubenswrapper[4814]: E0227 17:11:06.383442 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Feb 27 17:11:06 crc kubenswrapper[4814]: E0227 17:11:06.384730 4814 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9r85q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(c11159aa-3255-4574-81a3-182c8a30d970): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:11:06 crc kubenswrapper[4814]: E0227 17:11:06.387865 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="c11159aa-3255-4574-81a3-182c8a30d970" Feb 27 17:11:06 crc kubenswrapper[4814]: E0227 17:11:06.864369 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="c11159aa-3255-4574-81a3-182c8a30d970" Feb 27 17:11:21 crc kubenswrapper[4814]: I0227 17:11:21.419912 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 27 17:11:23 crc kubenswrapper[4814]: I0227 17:11:23.062339 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c11159aa-3255-4574-81a3-182c8a30d970","Type":"ContainerStarted","Data":"f837f3034c30072938725c0f2ba43c8dce3eb86064b075d33ec72f1e30337bcc"} Feb 27 17:11:23 crc kubenswrapper[4814]: I0227 17:11:23.082050 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.564957843 podStartE2EDuration="54.082018591s" podCreationTimestamp="2026-02-27 17:10:29 +0000 UTC" firstStartedPulling="2026-02-27 17:10:30.898917154 +0000 UTC m=+2843.351541974" lastFinishedPulling="2026-02-27 17:11:21.415977852 +0000 UTC m=+2893.868602722" observedRunningTime="2026-02-27 17:11:23.07815749 +0000 UTC m=+2895.530782360" watchObservedRunningTime="2026-02-27 17:11:23.082018591 +0000 UTC m=+2895.534643491" Feb 27 17:12:00 crc kubenswrapper[4814]: I0227 17:12:00.168784 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536872-mglb7"] Feb 27 17:12:00 crc kubenswrapper[4814]: I0227 17:12:00.172141 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536872-mglb7" Feb 27 17:12:00 crc kubenswrapper[4814]: I0227 17:12:00.179078 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:12:00 crc kubenswrapper[4814]: I0227 17:12:00.179170 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:12:00 crc kubenswrapper[4814]: I0227 17:12:00.179576 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:12:00 crc kubenswrapper[4814]: I0227 17:12:00.199714 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536872-mglb7"] Feb 27 17:12:00 crc kubenswrapper[4814]: I0227 17:12:00.214728 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn4n7\" (UniqueName: \"kubernetes.io/projected/dc246f97-cde3-494c-ad54-33ae57a84977-kube-api-access-zn4n7\") pod \"auto-csr-approver-29536872-mglb7\" (UID: \"dc246f97-cde3-494c-ad54-33ae57a84977\") " pod="openshift-infra/auto-csr-approver-29536872-mglb7" Feb 27 17:12:00 crc kubenswrapper[4814]: I0227 17:12:00.317502 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn4n7\" (UniqueName: \"kubernetes.io/projected/dc246f97-cde3-494c-ad54-33ae57a84977-kube-api-access-zn4n7\") pod \"auto-csr-approver-29536872-mglb7\" (UID: \"dc246f97-cde3-494c-ad54-33ae57a84977\") " pod="openshift-infra/auto-csr-approver-29536872-mglb7" Feb 27 17:12:00 crc kubenswrapper[4814]: I0227 17:12:00.351914 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn4n7\" (UniqueName: \"kubernetes.io/projected/dc246f97-cde3-494c-ad54-33ae57a84977-kube-api-access-zn4n7\") pod \"auto-csr-approver-29536872-mglb7\" (UID: \"dc246f97-cde3-494c-ad54-33ae57a84977\") " pod="openshift-infra/auto-csr-approver-29536872-mglb7" Feb 27 17:12:00 crc kubenswrapper[4814]: I0227 17:12:00.508479 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536872-mglb7" Feb 27 17:12:00 crc kubenswrapper[4814]: I0227 17:12:00.997782 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536872-mglb7"] Feb 27 17:12:01 crc kubenswrapper[4814]: W0227 17:12:01.010608 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc246f97_cde3_494c_ad54_33ae57a84977.slice/crio-2666c462bdb3ba827a6e3fc8eba3d9c27d9aaa49fd869ff56656f9c8ad6c4e26 WatchSource:0}: Error finding container 2666c462bdb3ba827a6e3fc8eba3d9c27d9aaa49fd869ff56656f9c8ad6c4e26: Status 404 returned error can't find the container with id 2666c462bdb3ba827a6e3fc8eba3d9c27d9aaa49fd869ff56656f9c8ad6c4e26 Feb 27 17:12:01 crc kubenswrapper[4814]: I0227 17:12:01.535470 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536872-mglb7" event={"ID":"dc246f97-cde3-494c-ad54-33ae57a84977","Type":"ContainerStarted","Data":"2666c462bdb3ba827a6e3fc8eba3d9c27d9aaa49fd869ff56656f9c8ad6c4e26"} Feb 27 17:12:02 crc kubenswrapper[4814]: I0227 17:12:02.547836 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536872-mglb7" event={"ID":"dc246f97-cde3-494c-ad54-33ae57a84977","Type":"ContainerStarted","Data":"0957a7790f278832919b9646e753694e88d630418eba1db7f87cd94639b2ce0c"} Feb 27 17:12:02 crc kubenswrapper[4814]: I0227 17:12:02.576995 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536872-mglb7" podStartSLOduration=1.479017711 podStartE2EDuration="2.57696717s" podCreationTimestamp="2026-02-27 17:12:00 +0000 UTC" firstStartedPulling="2026-02-27 17:12:01.016980305 +0000 UTC m=+2933.469605165" lastFinishedPulling="2026-02-27 17:12:02.114929784 +0000 UTC m=+2934.567554624" observedRunningTime="2026-02-27 17:12:02.57025106 +0000 UTC m=+2935.022875900" watchObservedRunningTime="2026-02-27 17:12:02.57696717 +0000 UTC m=+2935.029592040" Feb 27 17:12:03 crc kubenswrapper[4814]: I0227 17:12:03.563332 4814 generic.go:334] "Generic (PLEG): container finished" podID="dc246f97-cde3-494c-ad54-33ae57a84977" containerID="0957a7790f278832919b9646e753694e88d630418eba1db7f87cd94639b2ce0c" exitCode=0 Feb 27 17:12:03 crc kubenswrapper[4814]: I0227 17:12:03.563504 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536872-mglb7" event={"ID":"dc246f97-cde3-494c-ad54-33ae57a84977","Type":"ContainerDied","Data":"0957a7790f278832919b9646e753694e88d630418eba1db7f87cd94639b2ce0c"} Feb 27 17:12:04 crc kubenswrapper[4814]: I0227 17:12:04.974206 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536872-mglb7" Feb 27 17:12:05 crc kubenswrapper[4814]: I0227 17:12:05.116314 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn4n7\" (UniqueName: \"kubernetes.io/projected/dc246f97-cde3-494c-ad54-33ae57a84977-kube-api-access-zn4n7\") pod \"dc246f97-cde3-494c-ad54-33ae57a84977\" (UID: \"dc246f97-cde3-494c-ad54-33ae57a84977\") " Feb 27 17:12:05 crc kubenswrapper[4814]: I0227 17:12:05.144123 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc246f97-cde3-494c-ad54-33ae57a84977-kube-api-access-zn4n7" (OuterVolumeSpecName: "kube-api-access-zn4n7") pod "dc246f97-cde3-494c-ad54-33ae57a84977" (UID: "dc246f97-cde3-494c-ad54-33ae57a84977"). InnerVolumeSpecName "kube-api-access-zn4n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:12:05 crc kubenswrapper[4814]: I0227 17:12:05.220178 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn4n7\" (UniqueName: \"kubernetes.io/projected/dc246f97-cde3-494c-ad54-33ae57a84977-kube-api-access-zn4n7\") on node \"crc\" DevicePath \"\"" Feb 27 17:12:05 crc kubenswrapper[4814]: I0227 17:12:05.589894 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536872-mglb7" event={"ID":"dc246f97-cde3-494c-ad54-33ae57a84977","Type":"ContainerDied","Data":"2666c462bdb3ba827a6e3fc8eba3d9c27d9aaa49fd869ff56656f9c8ad6c4e26"} Feb 27 17:12:05 crc kubenswrapper[4814]: I0227 17:12:05.590090 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2666c462bdb3ba827a6e3fc8eba3d9c27d9aaa49fd869ff56656f9c8ad6c4e26" Feb 27 17:12:05 crc kubenswrapper[4814]: I0227 17:12:05.590143 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536872-mglb7" Feb 27 17:12:05 crc kubenswrapper[4814]: I0227 17:12:05.682444 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536866-qsflg"] Feb 27 17:12:05 crc kubenswrapper[4814]: I0227 17:12:05.686464 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536866-qsflg"] Feb 27 17:12:06 crc kubenswrapper[4814]: I0227 17:12:06.510229 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b18ebba8-0085-45f5-b3da-fa039307426c" path="/var/lib/kubelet/pods/b18ebba8-0085-45f5-b3da-fa039307426c/volumes" Feb 27 17:12:21 crc kubenswrapper[4814]: I0227 17:12:21.649531 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p6jgk"] Feb 27 17:12:21 crc kubenswrapper[4814]: E0227 17:12:21.650705 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc246f97-cde3-494c-ad54-33ae57a84977" containerName="oc" Feb 27 17:12:21 crc kubenswrapper[4814]: I0227 17:12:21.650722 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc246f97-cde3-494c-ad54-33ae57a84977" containerName="oc" Feb 27 17:12:21 crc kubenswrapper[4814]: I0227 17:12:21.650946 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc246f97-cde3-494c-ad54-33ae57a84977" containerName="oc" Feb 27 17:12:21 crc kubenswrapper[4814]: I0227 17:12:21.652644 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:21 crc kubenswrapper[4814]: I0227 17:12:21.671826 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p6jgk"] Feb 27 17:12:21 crc kubenswrapper[4814]: I0227 17:12:21.767802 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cae5c14d-0a23-4a11-8647-f52888751e31-catalog-content\") pod \"certified-operators-p6jgk\" (UID: \"cae5c14d-0a23-4a11-8647-f52888751e31\") " pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:21 crc kubenswrapper[4814]: I0227 17:12:21.767989 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cae5c14d-0a23-4a11-8647-f52888751e31-utilities\") pod \"certified-operators-p6jgk\" (UID: \"cae5c14d-0a23-4a11-8647-f52888751e31\") " pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:21 crc kubenswrapper[4814]: I0227 17:12:21.768373 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kql5s\" (UniqueName: \"kubernetes.io/projected/cae5c14d-0a23-4a11-8647-f52888751e31-kube-api-access-kql5s\") pod \"certified-operators-p6jgk\" (UID: \"cae5c14d-0a23-4a11-8647-f52888751e31\") " pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:21 crc kubenswrapper[4814]: I0227 17:12:21.870370 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cae5c14d-0a23-4a11-8647-f52888751e31-catalog-content\") pod \"certified-operators-p6jgk\" (UID: \"cae5c14d-0a23-4a11-8647-f52888751e31\") " pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:21 crc kubenswrapper[4814]: I0227 17:12:21.870816 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cae5c14d-0a23-4a11-8647-f52888751e31-utilities\") pod \"certified-operators-p6jgk\" (UID: \"cae5c14d-0a23-4a11-8647-f52888751e31\") " pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:21 crc kubenswrapper[4814]: I0227 17:12:21.871221 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cae5c14d-0a23-4a11-8647-f52888751e31-catalog-content\") pod \"certified-operators-p6jgk\" (UID: \"cae5c14d-0a23-4a11-8647-f52888751e31\") " pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:21 crc kubenswrapper[4814]: I0227 17:12:21.871556 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kql5s\" (UniqueName: \"kubernetes.io/projected/cae5c14d-0a23-4a11-8647-f52888751e31-kube-api-access-kql5s\") pod \"certified-operators-p6jgk\" (UID: \"cae5c14d-0a23-4a11-8647-f52888751e31\") " pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:21 crc kubenswrapper[4814]: I0227 17:12:21.871559 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cae5c14d-0a23-4a11-8647-f52888751e31-utilities\") pod \"certified-operators-p6jgk\" (UID: \"cae5c14d-0a23-4a11-8647-f52888751e31\") " pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:21 crc kubenswrapper[4814]: I0227 17:12:21.922780 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kql5s\" (UniqueName: \"kubernetes.io/projected/cae5c14d-0a23-4a11-8647-f52888751e31-kube-api-access-kql5s\") pod \"certified-operators-p6jgk\" (UID: \"cae5c14d-0a23-4a11-8647-f52888751e31\") " pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:21 crc kubenswrapper[4814]: I0227 17:12:21.983331 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:22 crc kubenswrapper[4814]: I0227 17:12:22.566876 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p6jgk"] Feb 27 17:12:22 crc kubenswrapper[4814]: I0227 17:12:22.902747 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:12:22 crc kubenswrapper[4814]: I0227 17:12:22.902865 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:12:23 crc kubenswrapper[4814]: I0227 17:12:23.337699 4814 generic.go:334] "Generic (PLEG): container finished" podID="cae5c14d-0a23-4a11-8647-f52888751e31" containerID="51579caf44c2b74436a32501f33d666f8b2e3691e6335d07477b354c9d86bd52" exitCode=0 Feb 27 17:12:23 crc kubenswrapper[4814]: I0227 17:12:23.337749 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p6jgk" event={"ID":"cae5c14d-0a23-4a11-8647-f52888751e31","Type":"ContainerDied","Data":"51579caf44c2b74436a32501f33d666f8b2e3691e6335d07477b354c9d86bd52"} Feb 27 17:12:23 crc kubenswrapper[4814]: I0227 17:12:23.338178 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p6jgk" event={"ID":"cae5c14d-0a23-4a11-8647-f52888751e31","Type":"ContainerStarted","Data":"a61bf6fcf94ac7427b1ce8b445372d31d2920720de1c963d7363ab3b3039ff71"} Feb 27 17:12:25 crc kubenswrapper[4814]: I0227 17:12:25.364154 4814 generic.go:334] "Generic (PLEG): container finished" podID="cae5c14d-0a23-4a11-8647-f52888751e31" containerID="ea3c05ababb4159afba169e2cf2c7e6c52333f6efd2b5e379ed1da5ef3d9a4fc" exitCode=0 Feb 27 17:12:25 crc kubenswrapper[4814]: I0227 17:12:25.364290 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p6jgk" event={"ID":"cae5c14d-0a23-4a11-8647-f52888751e31","Type":"ContainerDied","Data":"ea3c05ababb4159afba169e2cf2c7e6c52333f6efd2b5e379ed1da5ef3d9a4fc"} Feb 27 17:12:26 crc kubenswrapper[4814]: I0227 17:12:26.379488 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p6jgk" event={"ID":"cae5c14d-0a23-4a11-8647-f52888751e31","Type":"ContainerStarted","Data":"281105af4afc5f37ab44a7bf00c494332daa1cfc0b123af0be1d32443e2f464a"} Feb 27 17:12:26 crc kubenswrapper[4814]: I0227 17:12:26.417276 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p6jgk" podStartSLOduration=2.9073448109999998 podStartE2EDuration="5.417239101s" podCreationTimestamp="2026-02-27 17:12:21 +0000 UTC" firstStartedPulling="2026-02-27 17:12:23.339981424 +0000 UTC m=+2955.792606264" lastFinishedPulling="2026-02-27 17:12:25.849875694 +0000 UTC m=+2958.302500554" observedRunningTime="2026-02-27 17:12:26.399850929 +0000 UTC m=+2958.852475779" watchObservedRunningTime="2026-02-27 17:12:26.417239101 +0000 UTC m=+2958.869863941" Feb 27 17:12:31 crc kubenswrapper[4814]: I0227 17:12:31.990317 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:31 crc kubenswrapper[4814]: I0227 17:12:31.991009 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:32 crc kubenswrapper[4814]: I0227 17:12:32.047956 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:32 crc kubenswrapper[4814]: I0227 17:12:32.536041 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:32 crc kubenswrapper[4814]: I0227 17:12:32.589926 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p6jgk"] Feb 27 17:12:34 crc kubenswrapper[4814]: I0227 17:12:34.499215 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p6jgk" podUID="cae5c14d-0a23-4a11-8647-f52888751e31" containerName="registry-server" containerID="cri-o://281105af4afc5f37ab44a7bf00c494332daa1cfc0b123af0be1d32443e2f464a" gracePeriod=2 Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.083705 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.280922 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kql5s\" (UniqueName: \"kubernetes.io/projected/cae5c14d-0a23-4a11-8647-f52888751e31-kube-api-access-kql5s\") pod \"cae5c14d-0a23-4a11-8647-f52888751e31\" (UID: \"cae5c14d-0a23-4a11-8647-f52888751e31\") " Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.281457 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cae5c14d-0a23-4a11-8647-f52888751e31-catalog-content\") pod \"cae5c14d-0a23-4a11-8647-f52888751e31\" (UID: \"cae5c14d-0a23-4a11-8647-f52888751e31\") " Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.281683 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cae5c14d-0a23-4a11-8647-f52888751e31-utilities\") pod \"cae5c14d-0a23-4a11-8647-f52888751e31\" (UID: \"cae5c14d-0a23-4a11-8647-f52888751e31\") " Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.283531 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cae5c14d-0a23-4a11-8647-f52888751e31-utilities" (OuterVolumeSpecName: "utilities") pod "cae5c14d-0a23-4a11-8647-f52888751e31" (UID: "cae5c14d-0a23-4a11-8647-f52888751e31"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.291845 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cae5c14d-0a23-4a11-8647-f52888751e31-kube-api-access-kql5s" (OuterVolumeSpecName: "kube-api-access-kql5s") pod "cae5c14d-0a23-4a11-8647-f52888751e31" (UID: "cae5c14d-0a23-4a11-8647-f52888751e31"). InnerVolumeSpecName "kube-api-access-kql5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.353349 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cae5c14d-0a23-4a11-8647-f52888751e31-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cae5c14d-0a23-4a11-8647-f52888751e31" (UID: "cae5c14d-0a23-4a11-8647-f52888751e31"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.383839 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cae5c14d-0a23-4a11-8647-f52888751e31-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.383878 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kql5s\" (UniqueName: \"kubernetes.io/projected/cae5c14d-0a23-4a11-8647-f52888751e31-kube-api-access-kql5s\") on node \"crc\" DevicePath \"\"" Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.383892 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cae5c14d-0a23-4a11-8647-f52888751e31-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.516854 4814 generic.go:334] "Generic (PLEG): container finished" podID="cae5c14d-0a23-4a11-8647-f52888751e31" containerID="281105af4afc5f37ab44a7bf00c494332daa1cfc0b123af0be1d32443e2f464a" exitCode=0 Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.516919 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p6jgk" event={"ID":"cae5c14d-0a23-4a11-8647-f52888751e31","Type":"ContainerDied","Data":"281105af4afc5f37ab44a7bf00c494332daa1cfc0b123af0be1d32443e2f464a"} Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.516968 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p6jgk" Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.517005 4814 scope.go:117] "RemoveContainer" containerID="281105af4afc5f37ab44a7bf00c494332daa1cfc0b123af0be1d32443e2f464a" Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.516984 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p6jgk" event={"ID":"cae5c14d-0a23-4a11-8647-f52888751e31","Type":"ContainerDied","Data":"a61bf6fcf94ac7427b1ce8b445372d31d2920720de1c963d7363ab3b3039ff71"} Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.542393 4814 scope.go:117] "RemoveContainer" containerID="ea3c05ababb4159afba169e2cf2c7e6c52333f6efd2b5e379ed1da5ef3d9a4fc" Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.579352 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p6jgk"] Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.583617 4814 scope.go:117] "RemoveContainer" containerID="51579caf44c2b74436a32501f33d666f8b2e3691e6335d07477b354c9d86bd52" Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.594778 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p6jgk"] Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.646844 4814 scope.go:117] "RemoveContainer" containerID="281105af4afc5f37ab44a7bf00c494332daa1cfc0b123af0be1d32443e2f464a" Feb 27 17:12:35 crc kubenswrapper[4814]: E0227 17:12:35.647774 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"281105af4afc5f37ab44a7bf00c494332daa1cfc0b123af0be1d32443e2f464a\": container with ID starting with 281105af4afc5f37ab44a7bf00c494332daa1cfc0b123af0be1d32443e2f464a not found: ID does not exist" containerID="281105af4afc5f37ab44a7bf00c494332daa1cfc0b123af0be1d32443e2f464a" Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.647936 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"281105af4afc5f37ab44a7bf00c494332daa1cfc0b123af0be1d32443e2f464a"} err="failed to get container status \"281105af4afc5f37ab44a7bf00c494332daa1cfc0b123af0be1d32443e2f464a\": rpc error: code = NotFound desc = could not find container \"281105af4afc5f37ab44a7bf00c494332daa1cfc0b123af0be1d32443e2f464a\": container with ID starting with 281105af4afc5f37ab44a7bf00c494332daa1cfc0b123af0be1d32443e2f464a not found: ID does not exist" Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.648073 4814 scope.go:117] "RemoveContainer" containerID="ea3c05ababb4159afba169e2cf2c7e6c52333f6efd2b5e379ed1da5ef3d9a4fc" Feb 27 17:12:35 crc kubenswrapper[4814]: E0227 17:12:35.648836 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea3c05ababb4159afba169e2cf2c7e6c52333f6efd2b5e379ed1da5ef3d9a4fc\": container with ID starting with ea3c05ababb4159afba169e2cf2c7e6c52333f6efd2b5e379ed1da5ef3d9a4fc not found: ID does not exist" containerID="ea3c05ababb4159afba169e2cf2c7e6c52333f6efd2b5e379ed1da5ef3d9a4fc" Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.648977 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea3c05ababb4159afba169e2cf2c7e6c52333f6efd2b5e379ed1da5ef3d9a4fc"} err="failed to get container status \"ea3c05ababb4159afba169e2cf2c7e6c52333f6efd2b5e379ed1da5ef3d9a4fc\": rpc error: code = NotFound desc = could not find container \"ea3c05ababb4159afba169e2cf2c7e6c52333f6efd2b5e379ed1da5ef3d9a4fc\": container with ID starting with ea3c05ababb4159afba169e2cf2c7e6c52333f6efd2b5e379ed1da5ef3d9a4fc not found: ID does not exist" Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.649070 4814 scope.go:117] "RemoveContainer" containerID="51579caf44c2b74436a32501f33d666f8b2e3691e6335d07477b354c9d86bd52" Feb 27 17:12:35 crc kubenswrapper[4814]: E0227 17:12:35.649536 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51579caf44c2b74436a32501f33d666f8b2e3691e6335d07477b354c9d86bd52\": container with ID starting with 51579caf44c2b74436a32501f33d666f8b2e3691e6335d07477b354c9d86bd52 not found: ID does not exist" containerID="51579caf44c2b74436a32501f33d666f8b2e3691e6335d07477b354c9d86bd52" Feb 27 17:12:35 crc kubenswrapper[4814]: I0227 17:12:35.649627 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51579caf44c2b74436a32501f33d666f8b2e3691e6335d07477b354c9d86bd52"} err="failed to get container status \"51579caf44c2b74436a32501f33d666f8b2e3691e6335d07477b354c9d86bd52\": rpc error: code = NotFound desc = could not find container \"51579caf44c2b74436a32501f33d666f8b2e3691e6335d07477b354c9d86bd52\": container with ID starting with 51579caf44c2b74436a32501f33d666f8b2e3691e6335d07477b354c9d86bd52 not found: ID does not exist" Feb 27 17:12:36 crc kubenswrapper[4814]: I0227 17:12:36.507655 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cae5c14d-0a23-4a11-8647-f52888751e31" path="/var/lib/kubelet/pods/cae5c14d-0a23-4a11-8647-f52888751e31/volumes" Feb 27 17:12:52 crc kubenswrapper[4814]: I0227 17:12:52.902198 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:12:52 crc kubenswrapper[4814]: I0227 17:12:52.903121 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:13:06 crc kubenswrapper[4814]: I0227 17:13:06.430125 4814 scope.go:117] "RemoveContainer" containerID="043a5d2eb1d7bba4a9977ca8ca21b9bab9f0b44e4688076178e6ca5a0ff4157d" Feb 27 17:13:22 crc kubenswrapper[4814]: I0227 17:13:22.902641 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:13:22 crc kubenswrapper[4814]: I0227 17:13:22.903525 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:13:22 crc kubenswrapper[4814]: I0227 17:13:22.903588 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 17:13:22 crc kubenswrapper[4814]: I0227 17:13:22.904730 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"13fbd8981cbf65fbf38caf5962c43492d7821a7f2bb43648de09196463587955"} pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:13:22 crc kubenswrapper[4814]: I0227 17:13:22.904803 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" containerID="cri-o://13fbd8981cbf65fbf38caf5962c43492d7821a7f2bb43648de09196463587955" gracePeriod=600 Feb 27 17:13:23 crc kubenswrapper[4814]: I0227 17:13:23.151725 4814 generic.go:334] "Generic (PLEG): container finished" podID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerID="13fbd8981cbf65fbf38caf5962c43492d7821a7f2bb43648de09196463587955" exitCode=0 Feb 27 17:13:23 crc kubenswrapper[4814]: I0227 17:13:23.152195 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerDied","Data":"13fbd8981cbf65fbf38caf5962c43492d7821a7f2bb43648de09196463587955"} Feb 27 17:13:23 crc kubenswrapper[4814]: I0227 17:13:23.152244 4814 scope.go:117] "RemoveContainer" containerID="8992e22a0733a3b40d0435008d281d9ba98ea1c8ea0090f277f44f7af74e9e4a" Feb 27 17:13:24 crc kubenswrapper[4814]: I0227 17:13:24.168989 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f"} Feb 27 17:14:00 crc kubenswrapper[4814]: I0227 17:14:00.160355 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536874-jrrq4"] Feb 27 17:14:00 crc kubenswrapper[4814]: E0227 17:14:00.161300 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cae5c14d-0a23-4a11-8647-f52888751e31" containerName="registry-server" Feb 27 17:14:00 crc kubenswrapper[4814]: I0227 17:14:00.161315 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="cae5c14d-0a23-4a11-8647-f52888751e31" containerName="registry-server" Feb 27 17:14:00 crc kubenswrapper[4814]: E0227 17:14:00.161349 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cae5c14d-0a23-4a11-8647-f52888751e31" containerName="extract-utilities" Feb 27 17:14:00 crc kubenswrapper[4814]: I0227 17:14:00.161358 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="cae5c14d-0a23-4a11-8647-f52888751e31" containerName="extract-utilities" Feb 27 17:14:00 crc kubenswrapper[4814]: E0227 17:14:00.161379 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cae5c14d-0a23-4a11-8647-f52888751e31" containerName="extract-content" Feb 27 17:14:00 crc kubenswrapper[4814]: I0227 17:14:00.161387 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="cae5c14d-0a23-4a11-8647-f52888751e31" containerName="extract-content" Feb 27 17:14:00 crc kubenswrapper[4814]: I0227 17:14:00.161621 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="cae5c14d-0a23-4a11-8647-f52888751e31" containerName="registry-server" Feb 27 17:14:00 crc kubenswrapper[4814]: I0227 17:14:00.162310 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536874-jrrq4" Feb 27 17:14:00 crc kubenswrapper[4814]: I0227 17:14:00.165235 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:14:00 crc kubenswrapper[4814]: I0227 17:14:00.168520 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:14:00 crc kubenswrapper[4814]: I0227 17:14:00.168566 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:14:00 crc kubenswrapper[4814]: I0227 17:14:00.186500 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ct62\" (UniqueName: \"kubernetes.io/projected/a0f71995-4e4b-40f2-a0dd-c6203c1a07db-kube-api-access-7ct62\") pod \"auto-csr-approver-29536874-jrrq4\" (UID: \"a0f71995-4e4b-40f2-a0dd-c6203c1a07db\") " pod="openshift-infra/auto-csr-approver-29536874-jrrq4" Feb 27 17:14:00 crc kubenswrapper[4814]: I0227 17:14:00.192519 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536874-jrrq4"] Feb 27 17:14:00 crc kubenswrapper[4814]: I0227 17:14:00.288391 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ct62\" (UniqueName: \"kubernetes.io/projected/a0f71995-4e4b-40f2-a0dd-c6203c1a07db-kube-api-access-7ct62\") pod \"auto-csr-approver-29536874-jrrq4\" (UID: \"a0f71995-4e4b-40f2-a0dd-c6203c1a07db\") " pod="openshift-infra/auto-csr-approver-29536874-jrrq4" Feb 27 17:14:00 crc kubenswrapper[4814]: I0227 17:14:00.308670 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ct62\" (UniqueName: \"kubernetes.io/projected/a0f71995-4e4b-40f2-a0dd-c6203c1a07db-kube-api-access-7ct62\") pod \"auto-csr-approver-29536874-jrrq4\" (UID: \"a0f71995-4e4b-40f2-a0dd-c6203c1a07db\") " pod="openshift-infra/auto-csr-approver-29536874-jrrq4" Feb 27 17:14:00 crc kubenswrapper[4814]: I0227 17:14:00.497844 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536874-jrrq4" Feb 27 17:14:00 crc kubenswrapper[4814]: I0227 17:14:00.990613 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536874-jrrq4"] Feb 27 17:14:00 crc kubenswrapper[4814]: W0227 17:14:00.997684 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0f71995_4e4b_40f2_a0dd_c6203c1a07db.slice/crio-cb954df839921b84a02f4f75ee55c23c6a39b380a4db5debd096f1c4a8666963 WatchSource:0}: Error finding container cb954df839921b84a02f4f75ee55c23c6a39b380a4db5debd096f1c4a8666963: Status 404 returned error can't find the container with id cb954df839921b84a02f4f75ee55c23c6a39b380a4db5debd096f1c4a8666963 Feb 27 17:14:01 crc kubenswrapper[4814]: I0227 17:14:01.718130 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536874-jrrq4" event={"ID":"a0f71995-4e4b-40f2-a0dd-c6203c1a07db","Type":"ContainerStarted","Data":"cb954df839921b84a02f4f75ee55c23c6a39b380a4db5debd096f1c4a8666963"} Feb 27 17:14:02 crc kubenswrapper[4814]: I0227 17:14:02.729574 4814 generic.go:334] "Generic (PLEG): container finished" podID="a0f71995-4e4b-40f2-a0dd-c6203c1a07db" containerID="d92c35790902c9c4e0e7950ef8816c4abbedb6595b068f585567c75992780518" exitCode=0 Feb 27 17:14:02 crc kubenswrapper[4814]: I0227 17:14:02.729621 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536874-jrrq4" event={"ID":"a0f71995-4e4b-40f2-a0dd-c6203c1a07db","Type":"ContainerDied","Data":"d92c35790902c9c4e0e7950ef8816c4abbedb6595b068f585567c75992780518"} Feb 27 17:14:04 crc kubenswrapper[4814]: I0227 17:14:04.196593 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536874-jrrq4" Feb 27 17:14:04 crc kubenswrapper[4814]: I0227 17:14:04.287247 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ct62\" (UniqueName: \"kubernetes.io/projected/a0f71995-4e4b-40f2-a0dd-c6203c1a07db-kube-api-access-7ct62\") pod \"a0f71995-4e4b-40f2-a0dd-c6203c1a07db\" (UID: \"a0f71995-4e4b-40f2-a0dd-c6203c1a07db\") " Feb 27 17:14:04 crc kubenswrapper[4814]: I0227 17:14:04.304598 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0f71995-4e4b-40f2-a0dd-c6203c1a07db-kube-api-access-7ct62" (OuterVolumeSpecName: "kube-api-access-7ct62") pod "a0f71995-4e4b-40f2-a0dd-c6203c1a07db" (UID: "a0f71995-4e4b-40f2-a0dd-c6203c1a07db"). InnerVolumeSpecName "kube-api-access-7ct62". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:14:04 crc kubenswrapper[4814]: I0227 17:14:04.390614 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ct62\" (UniqueName: \"kubernetes.io/projected/a0f71995-4e4b-40f2-a0dd-c6203c1a07db-kube-api-access-7ct62\") on node \"crc\" DevicePath \"\"" Feb 27 17:14:04 crc kubenswrapper[4814]: I0227 17:14:04.751886 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536874-jrrq4" event={"ID":"a0f71995-4e4b-40f2-a0dd-c6203c1a07db","Type":"ContainerDied","Data":"cb954df839921b84a02f4f75ee55c23c6a39b380a4db5debd096f1c4a8666963"} Feb 27 17:14:04 crc kubenswrapper[4814]: I0227 17:14:04.751927 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb954df839921b84a02f4f75ee55c23c6a39b380a4db5debd096f1c4a8666963" Feb 27 17:14:04 crc kubenswrapper[4814]: I0227 17:14:04.751969 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536874-jrrq4" Feb 27 17:14:05 crc kubenswrapper[4814]: I0227 17:14:05.292787 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536868-8vvj4"] Feb 27 17:14:05 crc kubenswrapper[4814]: I0227 17:14:05.305790 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536868-8vvj4"] Feb 27 17:14:06 crc kubenswrapper[4814]: I0227 17:14:06.508752 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb327adc-9d58-4183-9e6e-866708bbae18" path="/var/lib/kubelet/pods/fb327adc-9d58-4183-9e6e-866708bbae18/volumes" Feb 27 17:14:06 crc kubenswrapper[4814]: I0227 17:14:06.545455 4814 scope.go:117] "RemoveContainer" containerID="873c09fc96c2432ed431fedafa4ebb9be4dcc6d0b95ee8eb9f59810ee5f03253" Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.181904 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d"] Feb 27 17:15:00 crc kubenswrapper[4814]: E0227 17:15:00.184738 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f71995-4e4b-40f2-a0dd-c6203c1a07db" containerName="oc" Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.184832 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f71995-4e4b-40f2-a0dd-c6203c1a07db" containerName="oc" Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.185178 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0f71995-4e4b-40f2-a0dd-c6203c1a07db" containerName="oc" Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.186399 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d" Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.189516 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.194249 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.202789 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d"] Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.289577 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd68v\" (UniqueName: \"kubernetes.io/projected/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-kube-api-access-vd68v\") pod \"collect-profiles-29536875-msn6d\" (UID: \"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d" Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.289687 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-secret-volume\") pod \"collect-profiles-29536875-msn6d\" (UID: \"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d" Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.289759 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-config-volume\") pod \"collect-profiles-29536875-msn6d\" (UID: \"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d" Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.391622 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-secret-volume\") pod \"collect-profiles-29536875-msn6d\" (UID: \"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d" Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.391730 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-config-volume\") pod \"collect-profiles-29536875-msn6d\" (UID: \"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d" Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.392013 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd68v\" (UniqueName: \"kubernetes.io/projected/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-kube-api-access-vd68v\") pod \"collect-profiles-29536875-msn6d\" (UID: \"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d" Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.393218 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-config-volume\") pod \"collect-profiles-29536875-msn6d\" (UID: \"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d" Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.399487 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-secret-volume\") pod \"collect-profiles-29536875-msn6d\" (UID: \"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d" Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.427286 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd68v\" (UniqueName: \"kubernetes.io/projected/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-kube-api-access-vd68v\") pod \"collect-profiles-29536875-msn6d\" (UID: \"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d" Feb 27 17:15:00 crc kubenswrapper[4814]: I0227 17:15:00.525330 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d" Feb 27 17:15:01 crc kubenswrapper[4814]: I0227 17:15:01.017385 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d"] Feb 27 17:15:01 crc kubenswrapper[4814]: I0227 17:15:01.683574 4814 generic.go:334] "Generic (PLEG): container finished" podID="ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2" containerID="29b0810e19076daa876c3434517793bb09f680ba9fdfd4667026876b905c8d3f" exitCode=0 Feb 27 17:15:01 crc kubenswrapper[4814]: I0227 17:15:01.683665 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d" event={"ID":"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2","Type":"ContainerDied","Data":"29b0810e19076daa876c3434517793bb09f680ba9fdfd4667026876b905c8d3f"} Feb 27 17:15:01 crc kubenswrapper[4814]: I0227 17:15:01.684121 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d" event={"ID":"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2","Type":"ContainerStarted","Data":"7473f05e0efb5f705016746fa821c09398ad7009b3c851ffb3991fee504d5c82"} Feb 27 17:15:03 crc kubenswrapper[4814]: I0227 17:15:03.146948 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d" Feb 27 17:15:03 crc kubenswrapper[4814]: I0227 17:15:03.258748 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd68v\" (UniqueName: \"kubernetes.io/projected/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-kube-api-access-vd68v\") pod \"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2\" (UID: \"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2\") " Feb 27 17:15:03 crc kubenswrapper[4814]: I0227 17:15:03.258887 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-config-volume\") pod \"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2\" (UID: \"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2\") " Feb 27 17:15:03 crc kubenswrapper[4814]: I0227 17:15:03.258938 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-secret-volume\") pod \"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2\" (UID: \"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2\") " Feb 27 17:15:03 crc kubenswrapper[4814]: I0227 17:15:03.259857 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-config-volume" (OuterVolumeSpecName: "config-volume") pod "ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2" (UID: "ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:15:03 crc kubenswrapper[4814]: I0227 17:15:03.271109 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-kube-api-access-vd68v" (OuterVolumeSpecName: "kube-api-access-vd68v") pod "ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2" (UID: "ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2"). InnerVolumeSpecName "kube-api-access-vd68v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:15:03 crc kubenswrapper[4814]: I0227 17:15:03.271113 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2" (UID: "ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:15:03 crc kubenswrapper[4814]: I0227 17:15:03.361410 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd68v\" (UniqueName: \"kubernetes.io/projected/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-kube-api-access-vd68v\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:03 crc kubenswrapper[4814]: I0227 17:15:03.361445 4814 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:03 crc kubenswrapper[4814]: I0227 17:15:03.361453 4814 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:03 crc kubenswrapper[4814]: I0227 17:15:03.714648 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d" event={"ID":"ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2","Type":"ContainerDied","Data":"7473f05e0efb5f705016746fa821c09398ad7009b3c851ffb3991fee504d5c82"} Feb 27 17:15:03 crc kubenswrapper[4814]: I0227 17:15:03.714707 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7473f05e0efb5f705016746fa821c09398ad7009b3c851ffb3991fee504d5c82" Feb 27 17:15:03 crc kubenswrapper[4814]: I0227 17:15:03.715196 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-msn6d" Feb 27 17:15:04 crc kubenswrapper[4814]: I0227 17:15:04.273219 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp"] Feb 27 17:15:04 crc kubenswrapper[4814]: I0227 17:15:04.286629 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536830-6dzbp"] Feb 27 17:15:04 crc kubenswrapper[4814]: I0227 17:15:04.527474 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6" path="/var/lib/kubelet/pods/1a75f56f-a4b8-413c-8bc0-b56f3eea4cc6/volumes" Feb 27 17:15:06 crc kubenswrapper[4814]: I0227 17:15:06.646452 4814 scope.go:117] "RemoveContainer" containerID="2dfa94b8890e8410ea578e0769f6b244b7eda1951bdf77e6e572a5b0c8b509fc" Feb 27 17:15:51 crc kubenswrapper[4814]: I0227 17:15:51.472553 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lpg6n"] Feb 27 17:15:51 crc kubenswrapper[4814]: E0227 17:15:51.473926 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2" containerName="collect-profiles" Feb 27 17:15:51 crc kubenswrapper[4814]: I0227 17:15:51.473948 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2" containerName="collect-profiles" Feb 27 17:15:51 crc kubenswrapper[4814]: I0227 17:15:51.474297 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab74a1a8-5cf4-4d00-8d20-251b7a3e45d2" containerName="collect-profiles" Feb 27 17:15:51 crc kubenswrapper[4814]: I0227 17:15:51.476425 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:15:51 crc kubenswrapper[4814]: I0227 17:15:51.493069 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lpg6n"] Feb 27 17:15:51 crc kubenswrapper[4814]: I0227 17:15:51.558018 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-catalog-content\") pod \"community-operators-lpg6n\" (UID: \"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb\") " pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:15:51 crc kubenswrapper[4814]: I0227 17:15:51.558129 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-utilities\") pod \"community-operators-lpg6n\" (UID: \"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb\") " pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:15:51 crc kubenswrapper[4814]: I0227 17:15:51.558603 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxvrh\" (UniqueName: \"kubernetes.io/projected/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-kube-api-access-cxvrh\") pod \"community-operators-lpg6n\" (UID: \"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb\") " pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:15:51 crc kubenswrapper[4814]: I0227 17:15:51.660581 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxvrh\" (UniqueName: \"kubernetes.io/projected/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-kube-api-access-cxvrh\") pod \"community-operators-lpg6n\" (UID: \"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb\") " pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:15:51 crc kubenswrapper[4814]: I0227 17:15:51.660719 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-catalog-content\") pod \"community-operators-lpg6n\" (UID: \"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb\") " pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:15:51 crc kubenswrapper[4814]: I0227 17:15:51.660757 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-utilities\") pod \"community-operators-lpg6n\" (UID: \"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb\") " pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:15:51 crc kubenswrapper[4814]: I0227 17:15:51.661310 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-catalog-content\") pod \"community-operators-lpg6n\" (UID: \"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb\") " pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:15:51 crc kubenswrapper[4814]: I0227 17:15:51.661461 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-utilities\") pod \"community-operators-lpg6n\" (UID: \"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb\") " pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:15:51 crc kubenswrapper[4814]: I0227 17:15:51.688184 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxvrh\" (UniqueName: \"kubernetes.io/projected/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-kube-api-access-cxvrh\") pod \"community-operators-lpg6n\" (UID: \"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb\") " pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:15:51 crc kubenswrapper[4814]: I0227 17:15:51.815446 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:15:52 crc kubenswrapper[4814]: I0227 17:15:52.363235 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lpg6n"] Feb 27 17:15:52 crc kubenswrapper[4814]: I0227 17:15:52.902065 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:15:52 crc kubenswrapper[4814]: I0227 17:15:52.902643 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:15:53 crc kubenswrapper[4814]: I0227 17:15:53.312657 4814 generic.go:334] "Generic (PLEG): container finished" podID="59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb" containerID="162261facc622626cb83fb71bcb879c459821f3bb2982d636b0f2f0e5db1f39e" exitCode=0 Feb 27 17:15:53 crc kubenswrapper[4814]: I0227 17:15:53.312714 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpg6n" event={"ID":"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb","Type":"ContainerDied","Data":"162261facc622626cb83fb71bcb879c459821f3bb2982d636b0f2f0e5db1f39e"} Feb 27 17:15:53 crc kubenswrapper[4814]: I0227 17:15:53.312751 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpg6n" event={"ID":"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb","Type":"ContainerStarted","Data":"1c81dfd27fc11e1f4e11c1f5e68d030879979a8f0b3968588f02a25d1dbacbf9"} Feb 27 17:15:53 crc kubenswrapper[4814]: I0227 17:15:53.315203 4814 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:15:54 crc kubenswrapper[4814]: I0227 17:15:54.325710 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpg6n" event={"ID":"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb","Type":"ContainerStarted","Data":"219218ee84953e5b8117cbdf1cdb363ab15ba264c121e7065a64a990bd55cbfc"} Feb 27 17:15:55 crc kubenswrapper[4814]: I0227 17:15:55.343849 4814 generic.go:334] "Generic (PLEG): container finished" podID="59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb" containerID="219218ee84953e5b8117cbdf1cdb363ab15ba264c121e7065a64a990bd55cbfc" exitCode=0 Feb 27 17:15:55 crc kubenswrapper[4814]: I0227 17:15:55.343993 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpg6n" event={"ID":"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb","Type":"ContainerDied","Data":"219218ee84953e5b8117cbdf1cdb363ab15ba264c121e7065a64a990bd55cbfc"} Feb 27 17:15:56 crc kubenswrapper[4814]: I0227 17:15:56.365392 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpg6n" event={"ID":"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb","Type":"ContainerStarted","Data":"b9422103df25ac5fc5193cbe12e47a7ae989ea36a9add113ce008378816088cc"} Feb 27 17:15:56 crc kubenswrapper[4814]: I0227 17:15:56.401975 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lpg6n" podStartSLOduration=2.680953786 podStartE2EDuration="5.401948373s" podCreationTimestamp="2026-02-27 17:15:51 +0000 UTC" firstStartedPulling="2026-02-27 17:15:53.314804688 +0000 UTC m=+3165.767429558" lastFinishedPulling="2026-02-27 17:15:56.035799275 +0000 UTC m=+3168.488424145" observedRunningTime="2026-02-27 17:15:56.397863375 +0000 UTC m=+3168.850488215" watchObservedRunningTime="2026-02-27 17:15:56.401948373 +0000 UTC m=+3168.854573213" Feb 27 17:16:00 crc kubenswrapper[4814]: I0227 17:16:00.178926 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536876-c6xqw"] Feb 27 17:16:00 crc kubenswrapper[4814]: I0227 17:16:00.183232 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536876-c6xqw" Feb 27 17:16:00 crc kubenswrapper[4814]: I0227 17:16:00.186658 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:16:00 crc kubenswrapper[4814]: I0227 17:16:00.186757 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:16:00 crc kubenswrapper[4814]: I0227 17:16:00.187242 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:16:00 crc kubenswrapper[4814]: I0227 17:16:00.201658 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536876-c6xqw"] Feb 27 17:16:00 crc kubenswrapper[4814]: I0227 17:16:00.377525 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rb7l\" (UniqueName: \"kubernetes.io/projected/8bfb3c95-d26c-4660-a3f7-208d65f53ea1-kube-api-access-9rb7l\") pod \"auto-csr-approver-29536876-c6xqw\" (UID: \"8bfb3c95-d26c-4660-a3f7-208d65f53ea1\") " pod="openshift-infra/auto-csr-approver-29536876-c6xqw" Feb 27 17:16:00 crc kubenswrapper[4814]: I0227 17:16:00.479232 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rb7l\" (UniqueName: \"kubernetes.io/projected/8bfb3c95-d26c-4660-a3f7-208d65f53ea1-kube-api-access-9rb7l\") pod \"auto-csr-approver-29536876-c6xqw\" (UID: \"8bfb3c95-d26c-4660-a3f7-208d65f53ea1\") " pod="openshift-infra/auto-csr-approver-29536876-c6xqw" Feb 27 17:16:00 crc kubenswrapper[4814]: I0227 17:16:00.503785 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rb7l\" (UniqueName: \"kubernetes.io/projected/8bfb3c95-d26c-4660-a3f7-208d65f53ea1-kube-api-access-9rb7l\") pod \"auto-csr-approver-29536876-c6xqw\" (UID: \"8bfb3c95-d26c-4660-a3f7-208d65f53ea1\") " pod="openshift-infra/auto-csr-approver-29536876-c6xqw" Feb 27 17:16:00 crc kubenswrapper[4814]: I0227 17:16:00.523425 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536876-c6xqw" Feb 27 17:16:01 crc kubenswrapper[4814]: I0227 17:16:01.072388 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536876-c6xqw"] Feb 27 17:16:01 crc kubenswrapper[4814]: W0227 17:16:01.082488 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bfb3c95_d26c_4660_a3f7_208d65f53ea1.slice/crio-5548c9d6bc83816aa914711592bb9058da59fbe004329b46dbd5bc738b29a2cd WatchSource:0}: Error finding container 5548c9d6bc83816aa914711592bb9058da59fbe004329b46dbd5bc738b29a2cd: Status 404 returned error can't find the container with id 5548c9d6bc83816aa914711592bb9058da59fbe004329b46dbd5bc738b29a2cd Feb 27 17:16:01 crc kubenswrapper[4814]: I0227 17:16:01.435898 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536876-c6xqw" event={"ID":"8bfb3c95-d26c-4660-a3f7-208d65f53ea1","Type":"ContainerStarted","Data":"5548c9d6bc83816aa914711592bb9058da59fbe004329b46dbd5bc738b29a2cd"} Feb 27 17:16:01 crc kubenswrapper[4814]: I0227 17:16:01.816742 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:16:01 crc kubenswrapper[4814]: I0227 17:16:01.816821 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:16:01 crc kubenswrapper[4814]: I0227 17:16:01.889141 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:16:02 crc kubenswrapper[4814]: I0227 17:16:02.523076 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:16:02 crc kubenswrapper[4814]: I0227 17:16:02.585336 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lpg6n"] Feb 27 17:16:03 crc kubenswrapper[4814]: I0227 17:16:03.476444 4814 generic.go:334] "Generic (PLEG): container finished" podID="8bfb3c95-d26c-4660-a3f7-208d65f53ea1" containerID="2e168673b962f2c6971291c455865f9e21b4b8800a761d0e3927a94a5a5a66c9" exitCode=0 Feb 27 17:16:03 crc kubenswrapper[4814]: I0227 17:16:03.477431 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536876-c6xqw" event={"ID":"8bfb3c95-d26c-4660-a3f7-208d65f53ea1","Type":"ContainerDied","Data":"2e168673b962f2c6971291c455865f9e21b4b8800a761d0e3927a94a5a5a66c9"} Feb 27 17:16:04 crc kubenswrapper[4814]: I0227 17:16:04.485125 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lpg6n" podUID="59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb" containerName="registry-server" containerID="cri-o://b9422103df25ac5fc5193cbe12e47a7ae989ea36a9add113ce008378816088cc" gracePeriod=2 Feb 27 17:16:04 crc kubenswrapper[4814]: I0227 17:16:04.887881 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536876-c6xqw" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.010953 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.090283 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rb7l\" (UniqueName: \"kubernetes.io/projected/8bfb3c95-d26c-4660-a3f7-208d65f53ea1-kube-api-access-9rb7l\") pod \"8bfb3c95-d26c-4660-a3f7-208d65f53ea1\" (UID: \"8bfb3c95-d26c-4660-a3f7-208d65f53ea1\") " Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.097855 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bfb3c95-d26c-4660-a3f7-208d65f53ea1-kube-api-access-9rb7l" (OuterVolumeSpecName: "kube-api-access-9rb7l") pod "8bfb3c95-d26c-4660-a3f7-208d65f53ea1" (UID: "8bfb3c95-d26c-4660-a3f7-208d65f53ea1"). InnerVolumeSpecName "kube-api-access-9rb7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.191545 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-utilities\") pod \"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb\" (UID: \"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb\") " Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.191590 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxvrh\" (UniqueName: \"kubernetes.io/projected/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-kube-api-access-cxvrh\") pod \"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb\" (UID: \"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb\") " Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.191744 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-catalog-content\") pod \"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb\" (UID: \"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb\") " Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.192062 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rb7l\" (UniqueName: \"kubernetes.io/projected/8bfb3c95-d26c-4660-a3f7-208d65f53ea1-kube-api-access-9rb7l\") on node \"crc\" DevicePath \"\"" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.192781 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-utilities" (OuterVolumeSpecName: "utilities") pod "59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb" (UID: "59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.196428 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-kube-api-access-cxvrh" (OuterVolumeSpecName: "kube-api-access-cxvrh") pod "59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb" (UID: "59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb"). InnerVolumeSpecName "kube-api-access-cxvrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.252654 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb" (UID: "59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.294446 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.294507 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.294527 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxvrh\" (UniqueName: \"kubernetes.io/projected/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb-kube-api-access-cxvrh\") on node \"crc\" DevicePath \"\"" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.499063 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536876-c6xqw" event={"ID":"8bfb3c95-d26c-4660-a3f7-208d65f53ea1","Type":"ContainerDied","Data":"5548c9d6bc83816aa914711592bb9058da59fbe004329b46dbd5bc738b29a2cd"} Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.499117 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5548c9d6bc83816aa914711592bb9058da59fbe004329b46dbd5bc738b29a2cd" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.499132 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536876-c6xqw" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.504049 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lpg6n" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.504151 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpg6n" event={"ID":"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb","Type":"ContainerDied","Data":"b9422103df25ac5fc5193cbe12e47a7ae989ea36a9add113ce008378816088cc"} Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.504242 4814 scope.go:117] "RemoveContainer" containerID="b9422103df25ac5fc5193cbe12e47a7ae989ea36a9add113ce008378816088cc" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.503922 4814 generic.go:334] "Generic (PLEG): container finished" podID="59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb" containerID="b9422103df25ac5fc5193cbe12e47a7ae989ea36a9add113ce008378816088cc" exitCode=0 Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.510052 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpg6n" event={"ID":"59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb","Type":"ContainerDied","Data":"1c81dfd27fc11e1f4e11c1f5e68d030879979a8f0b3968588f02a25d1dbacbf9"} Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.550514 4814 scope.go:117] "RemoveContainer" containerID="219218ee84953e5b8117cbdf1cdb363ab15ba264c121e7065a64a990bd55cbfc" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.574092 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lpg6n"] Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.584179 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lpg6n"] Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.594134 4814 scope.go:117] "RemoveContainer" containerID="162261facc622626cb83fb71bcb879c459821f3bb2982d636b0f2f0e5db1f39e" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.618292 4814 scope.go:117] "RemoveContainer" containerID="b9422103df25ac5fc5193cbe12e47a7ae989ea36a9add113ce008378816088cc" Feb 27 17:16:05 crc kubenswrapper[4814]: E0227 17:16:05.619581 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9422103df25ac5fc5193cbe12e47a7ae989ea36a9add113ce008378816088cc\": container with ID starting with b9422103df25ac5fc5193cbe12e47a7ae989ea36a9add113ce008378816088cc not found: ID does not exist" containerID="b9422103df25ac5fc5193cbe12e47a7ae989ea36a9add113ce008378816088cc" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.619633 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9422103df25ac5fc5193cbe12e47a7ae989ea36a9add113ce008378816088cc"} err="failed to get container status \"b9422103df25ac5fc5193cbe12e47a7ae989ea36a9add113ce008378816088cc\": rpc error: code = NotFound desc = could not find container \"b9422103df25ac5fc5193cbe12e47a7ae989ea36a9add113ce008378816088cc\": container with ID starting with b9422103df25ac5fc5193cbe12e47a7ae989ea36a9add113ce008378816088cc not found: ID does not exist" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.619665 4814 scope.go:117] "RemoveContainer" containerID="219218ee84953e5b8117cbdf1cdb363ab15ba264c121e7065a64a990bd55cbfc" Feb 27 17:16:05 crc kubenswrapper[4814]: E0227 17:16:05.620146 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"219218ee84953e5b8117cbdf1cdb363ab15ba264c121e7065a64a990bd55cbfc\": container with ID starting with 219218ee84953e5b8117cbdf1cdb363ab15ba264c121e7065a64a990bd55cbfc not found: ID does not exist" containerID="219218ee84953e5b8117cbdf1cdb363ab15ba264c121e7065a64a990bd55cbfc" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.620421 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"219218ee84953e5b8117cbdf1cdb363ab15ba264c121e7065a64a990bd55cbfc"} err="failed to get container status \"219218ee84953e5b8117cbdf1cdb363ab15ba264c121e7065a64a990bd55cbfc\": rpc error: code = NotFound desc = could not find container \"219218ee84953e5b8117cbdf1cdb363ab15ba264c121e7065a64a990bd55cbfc\": container with ID starting with 219218ee84953e5b8117cbdf1cdb363ab15ba264c121e7065a64a990bd55cbfc not found: ID does not exist" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.620582 4814 scope.go:117] "RemoveContainer" containerID="162261facc622626cb83fb71bcb879c459821f3bb2982d636b0f2f0e5db1f39e" Feb 27 17:16:05 crc kubenswrapper[4814]: E0227 17:16:05.621107 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"162261facc622626cb83fb71bcb879c459821f3bb2982d636b0f2f0e5db1f39e\": container with ID starting with 162261facc622626cb83fb71bcb879c459821f3bb2982d636b0f2f0e5db1f39e not found: ID does not exist" containerID="162261facc622626cb83fb71bcb879c459821f3bb2982d636b0f2f0e5db1f39e" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.621152 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"162261facc622626cb83fb71bcb879c459821f3bb2982d636b0f2f0e5db1f39e"} err="failed to get container status \"162261facc622626cb83fb71bcb879c459821f3bb2982d636b0f2f0e5db1f39e\": rpc error: code = NotFound desc = could not find container \"162261facc622626cb83fb71bcb879c459821f3bb2982d636b0f2f0e5db1f39e\": container with ID starting with 162261facc622626cb83fb71bcb879c459821f3bb2982d636b0f2f0e5db1f39e not found: ID does not exist" Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.980347 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536870-wdkzj"] Feb 27 17:16:05 crc kubenswrapper[4814]: I0227 17:16:05.992633 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536870-wdkzj"] Feb 27 17:16:06 crc kubenswrapper[4814]: I0227 17:16:06.506295 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33af9720-6802-4c2b-8fbf-37ec02ece573" path="/var/lib/kubelet/pods/33af9720-6802-4c2b-8fbf-37ec02ece573/volumes" Feb 27 17:16:06 crc kubenswrapper[4814]: I0227 17:16:06.507714 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb" path="/var/lib/kubelet/pods/59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb/volumes" Feb 27 17:16:06 crc kubenswrapper[4814]: I0227 17:16:06.751497 4814 scope.go:117] "RemoveContainer" containerID="e74711f48be7b2944ceec9791efefe4bbba67956665a3f3bc133bd0e81e84260" Feb 27 17:16:22 crc kubenswrapper[4814]: I0227 17:16:22.902736 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:16:22 crc kubenswrapper[4814]: I0227 17:16:22.903469 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:16:52 crc kubenswrapper[4814]: I0227 17:16:52.903082 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:16:52 crc kubenswrapper[4814]: I0227 17:16:52.904136 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:16:52 crc kubenswrapper[4814]: I0227 17:16:52.904238 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 17:16:52 crc kubenswrapper[4814]: I0227 17:16:52.906013 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f"} pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:16:52 crc kubenswrapper[4814]: I0227 17:16:52.906126 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" containerID="cri-o://422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" gracePeriod=600 Feb 27 17:16:53 crc kubenswrapper[4814]: E0227 17:16:53.073397 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:16:53 crc kubenswrapper[4814]: I0227 17:16:53.086759 4814 generic.go:334] "Generic (PLEG): container finished" podID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" exitCode=0 Feb 27 17:16:53 crc kubenswrapper[4814]: I0227 17:16:53.086829 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerDied","Data":"422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f"} Feb 27 17:16:53 crc kubenswrapper[4814]: I0227 17:16:53.086883 4814 scope.go:117] "RemoveContainer" containerID="13fbd8981cbf65fbf38caf5962c43492d7821a7f2bb43648de09196463587955" Feb 27 17:16:53 crc kubenswrapper[4814]: I0227 17:16:53.087949 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:16:53 crc kubenswrapper[4814]: E0227 17:16:53.088446 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:17:06 crc kubenswrapper[4814]: I0227 17:17:06.487430 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:17:06 crc kubenswrapper[4814]: E0227 17:17:06.488537 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.617821 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vzkzn"] Feb 27 17:17:16 crc kubenswrapper[4814]: E0227 17:17:16.618892 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bfb3c95-d26c-4660-a3f7-208d65f53ea1" containerName="oc" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.618911 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bfb3c95-d26c-4660-a3f7-208d65f53ea1" containerName="oc" Feb 27 17:17:16 crc kubenswrapper[4814]: E0227 17:17:16.618935 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb" containerName="extract-utilities" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.618943 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb" containerName="extract-utilities" Feb 27 17:17:16 crc kubenswrapper[4814]: E0227 17:17:16.618952 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb" containerName="registry-server" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.618960 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb" containerName="registry-server" Feb 27 17:17:16 crc kubenswrapper[4814]: E0227 17:17:16.618994 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb" containerName="extract-content" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.619002 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb" containerName="extract-content" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.619230 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="59018acf-0a8c-4fb0-a5a1-6fef5d69b1eb" containerName="registry-server" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.619244 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bfb3c95-d26c-4660-a3f7-208d65f53ea1" containerName="oc" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.621056 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.651755 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vzkzn"] Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.685774 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/050f7549-d916-4c3a-b811-e15ef5861639-catalog-content\") pod \"redhat-marketplace-vzkzn\" (UID: \"050f7549-d916-4c3a-b811-e15ef5861639\") " pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.685867 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/050f7549-d916-4c3a-b811-e15ef5861639-utilities\") pod \"redhat-marketplace-vzkzn\" (UID: \"050f7549-d916-4c3a-b811-e15ef5861639\") " pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.685923 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znvf7\" (UniqueName: \"kubernetes.io/projected/050f7549-d916-4c3a-b811-e15ef5861639-kube-api-access-znvf7\") pod \"redhat-marketplace-vzkzn\" (UID: \"050f7549-d916-4c3a-b811-e15ef5861639\") " pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.788040 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znvf7\" (UniqueName: \"kubernetes.io/projected/050f7549-d916-4c3a-b811-e15ef5861639-kube-api-access-znvf7\") pod \"redhat-marketplace-vzkzn\" (UID: \"050f7549-d916-4c3a-b811-e15ef5861639\") " pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.788281 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/050f7549-d916-4c3a-b811-e15ef5861639-catalog-content\") pod \"redhat-marketplace-vzkzn\" (UID: \"050f7549-d916-4c3a-b811-e15ef5861639\") " pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.788341 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/050f7549-d916-4c3a-b811-e15ef5861639-utilities\") pod \"redhat-marketplace-vzkzn\" (UID: \"050f7549-d916-4c3a-b811-e15ef5861639\") " pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.788750 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/050f7549-d916-4c3a-b811-e15ef5861639-catalog-content\") pod \"redhat-marketplace-vzkzn\" (UID: \"050f7549-d916-4c3a-b811-e15ef5861639\") " pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.788798 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/050f7549-d916-4c3a-b811-e15ef5861639-utilities\") pod \"redhat-marketplace-vzkzn\" (UID: \"050f7549-d916-4c3a-b811-e15ef5861639\") " pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.809931 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znvf7\" (UniqueName: \"kubernetes.io/projected/050f7549-d916-4c3a-b811-e15ef5861639-kube-api-access-znvf7\") pod \"redhat-marketplace-vzkzn\" (UID: \"050f7549-d916-4c3a-b811-e15ef5861639\") " pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:16 crc kubenswrapper[4814]: I0227 17:17:16.957159 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:17 crc kubenswrapper[4814]: I0227 17:17:17.441571 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vzkzn"] Feb 27 17:17:18 crc kubenswrapper[4814]: I0227 17:17:18.369071 4814 generic.go:334] "Generic (PLEG): container finished" podID="050f7549-d916-4c3a-b811-e15ef5861639" containerID="3c13c197e41ffee8f73c3467daa8379d3639b94ad8059b9fc394b005d018952a" exitCode=0 Feb 27 17:17:18 crc kubenswrapper[4814]: I0227 17:17:18.369154 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vzkzn" event={"ID":"050f7549-d916-4c3a-b811-e15ef5861639","Type":"ContainerDied","Data":"3c13c197e41ffee8f73c3467daa8379d3639b94ad8059b9fc394b005d018952a"} Feb 27 17:17:18 crc kubenswrapper[4814]: I0227 17:17:18.369461 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vzkzn" event={"ID":"050f7549-d916-4c3a-b811-e15ef5861639","Type":"ContainerStarted","Data":"907d067992eab21d3b06753c4ac90ad3dcea787172c5cdbe70fa1d5018ca053c"} Feb 27 17:17:18 crc kubenswrapper[4814]: I0227 17:17:18.498935 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:17:18 crc kubenswrapper[4814]: E0227 17:17:18.499744 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:17:19 crc kubenswrapper[4814]: I0227 17:17:19.384057 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vzkzn" event={"ID":"050f7549-d916-4c3a-b811-e15ef5861639","Type":"ContainerStarted","Data":"0dd9e9de5d0487430f008525e3228d8abf5bb66a0971174ef46b71e2146686b9"} Feb 27 17:17:20 crc kubenswrapper[4814]: I0227 17:17:20.398923 4814 generic.go:334] "Generic (PLEG): container finished" podID="050f7549-d916-4c3a-b811-e15ef5861639" containerID="0dd9e9de5d0487430f008525e3228d8abf5bb66a0971174ef46b71e2146686b9" exitCode=0 Feb 27 17:17:20 crc kubenswrapper[4814]: I0227 17:17:20.399095 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vzkzn" event={"ID":"050f7549-d916-4c3a-b811-e15ef5861639","Type":"ContainerDied","Data":"0dd9e9de5d0487430f008525e3228d8abf5bb66a0971174ef46b71e2146686b9"} Feb 27 17:17:21 crc kubenswrapper[4814]: I0227 17:17:21.422825 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vzkzn" event={"ID":"050f7549-d916-4c3a-b811-e15ef5861639","Type":"ContainerStarted","Data":"dc42e91fdb82c693b9341d47548f6c32a23830f1763ffcd8896c9e1fde2632d3"} Feb 27 17:17:21 crc kubenswrapper[4814]: I0227 17:17:21.452835 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vzkzn" podStartSLOduration=2.919388784 podStartE2EDuration="5.452807628s" podCreationTimestamp="2026-02-27 17:17:16 +0000 UTC" firstStartedPulling="2026-02-27 17:17:18.371888315 +0000 UTC m=+3250.824513185" lastFinishedPulling="2026-02-27 17:17:20.905307159 +0000 UTC m=+3253.357932029" observedRunningTime="2026-02-27 17:17:21.449547317 +0000 UTC m=+3253.902172157" watchObservedRunningTime="2026-02-27 17:17:21.452807628 +0000 UTC m=+3253.905432498" Feb 27 17:17:26 crc kubenswrapper[4814]: I0227 17:17:26.958391 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:26 crc kubenswrapper[4814]: I0227 17:17:26.959225 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:27 crc kubenswrapper[4814]: I0227 17:17:27.042048 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:27 crc kubenswrapper[4814]: I0227 17:17:27.554967 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:27 crc kubenswrapper[4814]: I0227 17:17:27.619624 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vzkzn"] Feb 27 17:17:29 crc kubenswrapper[4814]: I0227 17:17:29.512083 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vzkzn" podUID="050f7549-d916-4c3a-b811-e15ef5861639" containerName="registry-server" containerID="cri-o://dc42e91fdb82c693b9341d47548f6c32a23830f1763ffcd8896c9e1fde2632d3" gracePeriod=2 Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.064847 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.112297 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/050f7549-d916-4c3a-b811-e15ef5861639-catalog-content\") pod \"050f7549-d916-4c3a-b811-e15ef5861639\" (UID: \"050f7549-d916-4c3a-b811-e15ef5861639\") " Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.112529 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znvf7\" (UniqueName: \"kubernetes.io/projected/050f7549-d916-4c3a-b811-e15ef5861639-kube-api-access-znvf7\") pod \"050f7549-d916-4c3a-b811-e15ef5861639\" (UID: \"050f7549-d916-4c3a-b811-e15ef5861639\") " Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.112573 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/050f7549-d916-4c3a-b811-e15ef5861639-utilities\") pod \"050f7549-d916-4c3a-b811-e15ef5861639\" (UID: \"050f7549-d916-4c3a-b811-e15ef5861639\") " Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.113636 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/050f7549-d916-4c3a-b811-e15ef5861639-utilities" (OuterVolumeSpecName: "utilities") pod "050f7549-d916-4c3a-b811-e15ef5861639" (UID: "050f7549-d916-4c3a-b811-e15ef5861639"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.118030 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/050f7549-d916-4c3a-b811-e15ef5861639-kube-api-access-znvf7" (OuterVolumeSpecName: "kube-api-access-znvf7") pod "050f7549-d916-4c3a-b811-e15ef5861639" (UID: "050f7549-d916-4c3a-b811-e15ef5861639"). InnerVolumeSpecName "kube-api-access-znvf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.142283 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/050f7549-d916-4c3a-b811-e15ef5861639-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "050f7549-d916-4c3a-b811-e15ef5861639" (UID: "050f7549-d916-4c3a-b811-e15ef5861639"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.215448 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/050f7549-d916-4c3a-b811-e15ef5861639-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.215479 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znvf7\" (UniqueName: \"kubernetes.io/projected/050f7549-d916-4c3a-b811-e15ef5861639-kube-api-access-znvf7\") on node \"crc\" DevicePath \"\"" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.215493 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/050f7549-d916-4c3a-b811-e15ef5861639-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.494735 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:17:30 crc kubenswrapper[4814]: E0227 17:17:30.495583 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.529984 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vzkzn" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.530058 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vzkzn" event={"ID":"050f7549-d916-4c3a-b811-e15ef5861639","Type":"ContainerDied","Data":"dc42e91fdb82c693b9341d47548f6c32a23830f1763ffcd8896c9e1fde2632d3"} Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.530169 4814 scope.go:117] "RemoveContainer" containerID="dc42e91fdb82c693b9341d47548f6c32a23830f1763ffcd8896c9e1fde2632d3" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.530921 4814 generic.go:334] "Generic (PLEG): container finished" podID="050f7549-d916-4c3a-b811-e15ef5861639" containerID="dc42e91fdb82c693b9341d47548f6c32a23830f1763ffcd8896c9e1fde2632d3" exitCode=0 Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.530997 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vzkzn" event={"ID":"050f7549-d916-4c3a-b811-e15ef5861639","Type":"ContainerDied","Data":"907d067992eab21d3b06753c4ac90ad3dcea787172c5cdbe70fa1d5018ca053c"} Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.569913 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vzkzn"] Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.581206 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vzkzn"] Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.581821 4814 scope.go:117] "RemoveContainer" containerID="0dd9e9de5d0487430f008525e3228d8abf5bb66a0971174ef46b71e2146686b9" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.607168 4814 scope.go:117] "RemoveContainer" containerID="3c13c197e41ffee8f73c3467daa8379d3639b94ad8059b9fc394b005d018952a" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.665536 4814 scope.go:117] "RemoveContainer" containerID="dc42e91fdb82c693b9341d47548f6c32a23830f1763ffcd8896c9e1fde2632d3" Feb 27 17:17:30 crc kubenswrapper[4814]: E0227 17:17:30.666157 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc42e91fdb82c693b9341d47548f6c32a23830f1763ffcd8896c9e1fde2632d3\": container with ID starting with dc42e91fdb82c693b9341d47548f6c32a23830f1763ffcd8896c9e1fde2632d3 not found: ID does not exist" containerID="dc42e91fdb82c693b9341d47548f6c32a23830f1763ffcd8896c9e1fde2632d3" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.666205 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc42e91fdb82c693b9341d47548f6c32a23830f1763ffcd8896c9e1fde2632d3"} err="failed to get container status \"dc42e91fdb82c693b9341d47548f6c32a23830f1763ffcd8896c9e1fde2632d3\": rpc error: code = NotFound desc = could not find container \"dc42e91fdb82c693b9341d47548f6c32a23830f1763ffcd8896c9e1fde2632d3\": container with ID starting with dc42e91fdb82c693b9341d47548f6c32a23830f1763ffcd8896c9e1fde2632d3 not found: ID does not exist" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.666239 4814 scope.go:117] "RemoveContainer" containerID="0dd9e9de5d0487430f008525e3228d8abf5bb66a0971174ef46b71e2146686b9" Feb 27 17:17:30 crc kubenswrapper[4814]: E0227 17:17:30.666780 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dd9e9de5d0487430f008525e3228d8abf5bb66a0971174ef46b71e2146686b9\": container with ID starting with 0dd9e9de5d0487430f008525e3228d8abf5bb66a0971174ef46b71e2146686b9 not found: ID does not exist" containerID="0dd9e9de5d0487430f008525e3228d8abf5bb66a0971174ef46b71e2146686b9" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.666808 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dd9e9de5d0487430f008525e3228d8abf5bb66a0971174ef46b71e2146686b9"} err="failed to get container status \"0dd9e9de5d0487430f008525e3228d8abf5bb66a0971174ef46b71e2146686b9\": rpc error: code = NotFound desc = could not find container \"0dd9e9de5d0487430f008525e3228d8abf5bb66a0971174ef46b71e2146686b9\": container with ID starting with 0dd9e9de5d0487430f008525e3228d8abf5bb66a0971174ef46b71e2146686b9 not found: ID does not exist" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.666827 4814 scope.go:117] "RemoveContainer" containerID="3c13c197e41ffee8f73c3467daa8379d3639b94ad8059b9fc394b005d018952a" Feb 27 17:17:30 crc kubenswrapper[4814]: E0227 17:17:30.667132 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c13c197e41ffee8f73c3467daa8379d3639b94ad8059b9fc394b005d018952a\": container with ID starting with 3c13c197e41ffee8f73c3467daa8379d3639b94ad8059b9fc394b005d018952a not found: ID does not exist" containerID="3c13c197e41ffee8f73c3467daa8379d3639b94ad8059b9fc394b005d018952a" Feb 27 17:17:30 crc kubenswrapper[4814]: I0227 17:17:30.667167 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c13c197e41ffee8f73c3467daa8379d3639b94ad8059b9fc394b005d018952a"} err="failed to get container status \"3c13c197e41ffee8f73c3467daa8379d3639b94ad8059b9fc394b005d018952a\": rpc error: code = NotFound desc = could not find container \"3c13c197e41ffee8f73c3467daa8379d3639b94ad8059b9fc394b005d018952a\": container with ID starting with 3c13c197e41ffee8f73c3467daa8379d3639b94ad8059b9fc394b005d018952a not found: ID does not exist" Feb 27 17:17:32 crc kubenswrapper[4814]: I0227 17:17:32.507487 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="050f7549-d916-4c3a-b811-e15ef5861639" path="/var/lib/kubelet/pods/050f7549-d916-4c3a-b811-e15ef5861639/volumes" Feb 27 17:17:43 crc kubenswrapper[4814]: I0227 17:17:43.487339 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:17:43 crc kubenswrapper[4814]: E0227 17:17:43.487978 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:17:54 crc kubenswrapper[4814]: I0227 17:17:54.488401 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:17:54 crc kubenswrapper[4814]: E0227 17:17:54.489443 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:18:00 crc kubenswrapper[4814]: I0227 17:18:00.205362 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536878-qb7qq"] Feb 27 17:18:00 crc kubenswrapper[4814]: E0227 17:18:00.206129 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="050f7549-d916-4c3a-b811-e15ef5861639" containerName="registry-server" Feb 27 17:18:00 crc kubenswrapper[4814]: I0227 17:18:00.206142 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="050f7549-d916-4c3a-b811-e15ef5861639" containerName="registry-server" Feb 27 17:18:00 crc kubenswrapper[4814]: E0227 17:18:00.206155 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="050f7549-d916-4c3a-b811-e15ef5861639" containerName="extract-utilities" Feb 27 17:18:00 crc kubenswrapper[4814]: I0227 17:18:00.206161 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="050f7549-d916-4c3a-b811-e15ef5861639" containerName="extract-utilities" Feb 27 17:18:00 crc kubenswrapper[4814]: E0227 17:18:00.206176 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="050f7549-d916-4c3a-b811-e15ef5861639" containerName="extract-content" Feb 27 17:18:00 crc kubenswrapper[4814]: I0227 17:18:00.206183 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="050f7549-d916-4c3a-b811-e15ef5861639" containerName="extract-content" Feb 27 17:18:00 crc kubenswrapper[4814]: I0227 17:18:00.206386 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="050f7549-d916-4c3a-b811-e15ef5861639" containerName="registry-server" Feb 27 17:18:00 crc kubenswrapper[4814]: I0227 17:18:00.206981 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536878-qb7qq" Feb 27 17:18:00 crc kubenswrapper[4814]: I0227 17:18:00.209304 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:18:00 crc kubenswrapper[4814]: I0227 17:18:00.212918 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:18:00 crc kubenswrapper[4814]: I0227 17:18:00.221183 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:18:00 crc kubenswrapper[4814]: I0227 17:18:00.224457 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536878-qb7qq"] Feb 27 17:18:00 crc kubenswrapper[4814]: I0227 17:18:00.306688 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x57gv\" (UniqueName: \"kubernetes.io/projected/bdc6e071-8772-4c9e-811e-f2c747136f26-kube-api-access-x57gv\") pod \"auto-csr-approver-29536878-qb7qq\" (UID: \"bdc6e071-8772-4c9e-811e-f2c747136f26\") " pod="openshift-infra/auto-csr-approver-29536878-qb7qq" Feb 27 17:18:00 crc kubenswrapper[4814]: I0227 17:18:00.409050 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x57gv\" (UniqueName: \"kubernetes.io/projected/bdc6e071-8772-4c9e-811e-f2c747136f26-kube-api-access-x57gv\") pod \"auto-csr-approver-29536878-qb7qq\" (UID: \"bdc6e071-8772-4c9e-811e-f2c747136f26\") " pod="openshift-infra/auto-csr-approver-29536878-qb7qq" Feb 27 17:18:00 crc kubenswrapper[4814]: I0227 17:18:00.434107 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x57gv\" (UniqueName: \"kubernetes.io/projected/bdc6e071-8772-4c9e-811e-f2c747136f26-kube-api-access-x57gv\") pod \"auto-csr-approver-29536878-qb7qq\" (UID: \"bdc6e071-8772-4c9e-811e-f2c747136f26\") " pod="openshift-infra/auto-csr-approver-29536878-qb7qq" Feb 27 17:18:00 crc kubenswrapper[4814]: I0227 17:18:00.530529 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536878-qb7qq" Feb 27 17:18:00 crc kubenswrapper[4814]: I0227 17:18:00.989787 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536878-qb7qq"] Feb 27 17:18:01 crc kubenswrapper[4814]: I0227 17:18:01.879393 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536878-qb7qq" event={"ID":"bdc6e071-8772-4c9e-811e-f2c747136f26","Type":"ContainerStarted","Data":"e23e41dc28f195711a6cbc58bf0420c06eeb9110f2d723a68f5d8373e45adb41"} Feb 27 17:18:02 crc kubenswrapper[4814]: I0227 17:18:02.891360 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536878-qb7qq" event={"ID":"bdc6e071-8772-4c9e-811e-f2c747136f26","Type":"ContainerStarted","Data":"e6727f389a8f180b33335a320d5910e87674b4107164d84d2e352b0b35e6e40d"} Feb 27 17:18:02 crc kubenswrapper[4814]: I0227 17:18:02.917881 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536878-qb7qq" podStartSLOduration=1.459656817 podStartE2EDuration="2.91785682s" podCreationTimestamp="2026-02-27 17:18:00 +0000 UTC" firstStartedPulling="2026-02-27 17:18:00.986417093 +0000 UTC m=+3293.439041953" lastFinishedPulling="2026-02-27 17:18:02.444617136 +0000 UTC m=+3294.897241956" observedRunningTime="2026-02-27 17:18:02.914840316 +0000 UTC m=+3295.367465156" watchObservedRunningTime="2026-02-27 17:18:02.91785682 +0000 UTC m=+3295.370481660" Feb 27 17:18:03 crc kubenswrapper[4814]: I0227 17:18:03.905070 4814 generic.go:334] "Generic (PLEG): container finished" podID="bdc6e071-8772-4c9e-811e-f2c747136f26" containerID="e6727f389a8f180b33335a320d5910e87674b4107164d84d2e352b0b35e6e40d" exitCode=0 Feb 27 17:18:03 crc kubenswrapper[4814]: I0227 17:18:03.905118 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536878-qb7qq" event={"ID":"bdc6e071-8772-4c9e-811e-f2c747136f26","Type":"ContainerDied","Data":"e6727f389a8f180b33335a320d5910e87674b4107164d84d2e352b0b35e6e40d"} Feb 27 17:18:05 crc kubenswrapper[4814]: I0227 17:18:05.359208 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536878-qb7qq" Feb 27 17:18:05 crc kubenswrapper[4814]: I0227 17:18:05.409809 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x57gv\" (UniqueName: \"kubernetes.io/projected/bdc6e071-8772-4c9e-811e-f2c747136f26-kube-api-access-x57gv\") pod \"bdc6e071-8772-4c9e-811e-f2c747136f26\" (UID: \"bdc6e071-8772-4c9e-811e-f2c747136f26\") " Feb 27 17:18:05 crc kubenswrapper[4814]: I0227 17:18:05.416526 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdc6e071-8772-4c9e-811e-f2c747136f26-kube-api-access-x57gv" (OuterVolumeSpecName: "kube-api-access-x57gv") pod "bdc6e071-8772-4c9e-811e-f2c747136f26" (UID: "bdc6e071-8772-4c9e-811e-f2c747136f26"). InnerVolumeSpecName "kube-api-access-x57gv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:18:05 crc kubenswrapper[4814]: I0227 17:18:05.512575 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x57gv\" (UniqueName: \"kubernetes.io/projected/bdc6e071-8772-4c9e-811e-f2c747136f26-kube-api-access-x57gv\") on node \"crc\" DevicePath \"\"" Feb 27 17:18:05 crc kubenswrapper[4814]: I0227 17:18:05.944233 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536878-qb7qq" event={"ID":"bdc6e071-8772-4c9e-811e-f2c747136f26","Type":"ContainerDied","Data":"e23e41dc28f195711a6cbc58bf0420c06eeb9110f2d723a68f5d8373e45adb41"} Feb 27 17:18:05 crc kubenswrapper[4814]: I0227 17:18:05.944355 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e23e41dc28f195711a6cbc58bf0420c06eeb9110f2d723a68f5d8373e45adb41" Feb 27 17:18:05 crc kubenswrapper[4814]: I0227 17:18:05.944499 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536878-qb7qq" Feb 27 17:18:06 crc kubenswrapper[4814]: I0227 17:18:06.004722 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536872-mglb7"] Feb 27 17:18:06 crc kubenswrapper[4814]: I0227 17:18:06.013078 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536872-mglb7"] Feb 27 17:18:06 crc kubenswrapper[4814]: I0227 17:18:06.487723 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:18:06 crc kubenswrapper[4814]: E0227 17:18:06.488013 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:18:06 crc kubenswrapper[4814]: I0227 17:18:06.496836 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc246f97-cde3-494c-ad54-33ae57a84977" path="/var/lib/kubelet/pods/dc246f97-cde3-494c-ad54-33ae57a84977/volumes" Feb 27 17:18:06 crc kubenswrapper[4814]: I0227 17:18:06.882892 4814 scope.go:117] "RemoveContainer" containerID="0957a7790f278832919b9646e753694e88d630418eba1db7f87cd94639b2ce0c" Feb 27 17:18:17 crc kubenswrapper[4814]: I0227 17:18:17.488352 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:18:17 crc kubenswrapper[4814]: E0227 17:18:17.489473 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:18:31 crc kubenswrapper[4814]: I0227 17:18:31.488161 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:18:31 crc kubenswrapper[4814]: E0227 17:18:31.490108 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:18:44 crc kubenswrapper[4814]: I0227 17:18:44.488195 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:18:44 crc kubenswrapper[4814]: E0227 17:18:44.489481 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:18:56 crc kubenswrapper[4814]: I0227 17:18:56.488226 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:18:56 crc kubenswrapper[4814]: E0227 17:18:56.489109 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:19:07 crc kubenswrapper[4814]: I0227 17:19:07.487275 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:19:07 crc kubenswrapper[4814]: E0227 17:19:07.488090 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:19:19 crc kubenswrapper[4814]: I0227 17:19:19.488194 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:19:19 crc kubenswrapper[4814]: E0227 17:19:19.489171 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:19:33 crc kubenswrapper[4814]: I0227 17:19:33.487598 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:19:33 crc kubenswrapper[4814]: E0227 17:19:33.488682 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:19:48 crc kubenswrapper[4814]: I0227 17:19:48.496934 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:19:48 crc kubenswrapper[4814]: E0227 17:19:48.498028 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:20:00 crc kubenswrapper[4814]: I0227 17:20:00.156778 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536880-pbzpg"] Feb 27 17:20:00 crc kubenswrapper[4814]: E0227 17:20:00.157858 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdc6e071-8772-4c9e-811e-f2c747136f26" containerName="oc" Feb 27 17:20:00 crc kubenswrapper[4814]: I0227 17:20:00.157878 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdc6e071-8772-4c9e-811e-f2c747136f26" containerName="oc" Feb 27 17:20:00 crc kubenswrapper[4814]: I0227 17:20:00.158197 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdc6e071-8772-4c9e-811e-f2c747136f26" containerName="oc" Feb 27 17:20:00 crc kubenswrapper[4814]: I0227 17:20:00.159173 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536880-pbzpg" Feb 27 17:20:00 crc kubenswrapper[4814]: I0227 17:20:00.161727 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:20:00 crc kubenswrapper[4814]: I0227 17:20:00.163189 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:20:00 crc kubenswrapper[4814]: I0227 17:20:00.163605 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:20:00 crc kubenswrapper[4814]: I0227 17:20:00.165809 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536880-pbzpg"] Feb 27 17:20:00 crc kubenswrapper[4814]: I0227 17:20:00.217724 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gvjg\" (UniqueName: \"kubernetes.io/projected/42a61257-51ec-49cd-9e0f-3f872cc8a93b-kube-api-access-8gvjg\") pod \"auto-csr-approver-29536880-pbzpg\" (UID: \"42a61257-51ec-49cd-9e0f-3f872cc8a93b\") " pod="openshift-infra/auto-csr-approver-29536880-pbzpg" Feb 27 17:20:00 crc kubenswrapper[4814]: I0227 17:20:00.320405 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gvjg\" (UniqueName: \"kubernetes.io/projected/42a61257-51ec-49cd-9e0f-3f872cc8a93b-kube-api-access-8gvjg\") pod \"auto-csr-approver-29536880-pbzpg\" (UID: \"42a61257-51ec-49cd-9e0f-3f872cc8a93b\") " pod="openshift-infra/auto-csr-approver-29536880-pbzpg" Feb 27 17:20:00 crc kubenswrapper[4814]: I0227 17:20:00.351810 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gvjg\" (UniqueName: \"kubernetes.io/projected/42a61257-51ec-49cd-9e0f-3f872cc8a93b-kube-api-access-8gvjg\") pod \"auto-csr-approver-29536880-pbzpg\" (UID: \"42a61257-51ec-49cd-9e0f-3f872cc8a93b\") " pod="openshift-infra/auto-csr-approver-29536880-pbzpg" Feb 27 17:20:00 crc kubenswrapper[4814]: I0227 17:20:00.495700 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536880-pbzpg" Feb 27 17:20:00 crc kubenswrapper[4814]: I0227 17:20:00.995643 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536880-pbzpg"] Feb 27 17:20:01 crc kubenswrapper[4814]: I0227 17:20:01.254222 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536880-pbzpg" event={"ID":"42a61257-51ec-49cd-9e0f-3f872cc8a93b","Type":"ContainerStarted","Data":"b5ac831712ed002b86ca5a09281303159c957f755e78966f627ff790ace6d33d"} Feb 27 17:20:01 crc kubenswrapper[4814]: I0227 17:20:01.488748 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:20:01 crc kubenswrapper[4814]: E0227 17:20:01.489346 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:20:03 crc kubenswrapper[4814]: I0227 17:20:03.281811 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536880-pbzpg" event={"ID":"42a61257-51ec-49cd-9e0f-3f872cc8a93b","Type":"ContainerDied","Data":"de34d4f24044ae6e3fb99b5c983f19f17262db1f14fdb116f3cb03b0c052bbd1"} Feb 27 17:20:03 crc kubenswrapper[4814]: I0227 17:20:03.281686 4814 generic.go:334] "Generic (PLEG): container finished" podID="42a61257-51ec-49cd-9e0f-3f872cc8a93b" containerID="de34d4f24044ae6e3fb99b5c983f19f17262db1f14fdb116f3cb03b0c052bbd1" exitCode=0 Feb 27 17:20:04 crc kubenswrapper[4814]: I0227 17:20:04.782221 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536880-pbzpg" Feb 27 17:20:04 crc kubenswrapper[4814]: I0227 17:20:04.927811 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gvjg\" (UniqueName: \"kubernetes.io/projected/42a61257-51ec-49cd-9e0f-3f872cc8a93b-kube-api-access-8gvjg\") pod \"42a61257-51ec-49cd-9e0f-3f872cc8a93b\" (UID: \"42a61257-51ec-49cd-9e0f-3f872cc8a93b\") " Feb 27 17:20:04 crc kubenswrapper[4814]: I0227 17:20:04.934446 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42a61257-51ec-49cd-9e0f-3f872cc8a93b-kube-api-access-8gvjg" (OuterVolumeSpecName: "kube-api-access-8gvjg") pod "42a61257-51ec-49cd-9e0f-3f872cc8a93b" (UID: "42a61257-51ec-49cd-9e0f-3f872cc8a93b"). InnerVolumeSpecName "kube-api-access-8gvjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:05 crc kubenswrapper[4814]: I0227 17:20:05.030659 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gvjg\" (UniqueName: \"kubernetes.io/projected/42a61257-51ec-49cd-9e0f-3f872cc8a93b-kube-api-access-8gvjg\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:05 crc kubenswrapper[4814]: I0227 17:20:05.307879 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536880-pbzpg" event={"ID":"42a61257-51ec-49cd-9e0f-3f872cc8a93b","Type":"ContainerDied","Data":"b5ac831712ed002b86ca5a09281303159c957f755e78966f627ff790ace6d33d"} Feb 27 17:20:05 crc kubenswrapper[4814]: I0227 17:20:05.307932 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5ac831712ed002b86ca5a09281303159c957f755e78966f627ff790ace6d33d" Feb 27 17:20:05 crc kubenswrapper[4814]: I0227 17:20:05.308434 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536880-pbzpg" Feb 27 17:20:05 crc kubenswrapper[4814]: I0227 17:20:05.873608 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536874-jrrq4"] Feb 27 17:20:05 crc kubenswrapper[4814]: I0227 17:20:05.885301 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536874-jrrq4"] Feb 27 17:20:06 crc kubenswrapper[4814]: I0227 17:20:06.506715 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0f71995-4e4b-40f2-a0dd-c6203c1a07db" path="/var/lib/kubelet/pods/a0f71995-4e4b-40f2-a0dd-c6203c1a07db/volumes" Feb 27 17:20:07 crc kubenswrapper[4814]: I0227 17:20:07.008027 4814 scope.go:117] "RemoveContainer" containerID="d92c35790902c9c4e0e7950ef8816c4abbedb6595b068f585567c75992780518" Feb 27 17:20:13 crc kubenswrapper[4814]: I0227 17:20:13.487545 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:20:13 crc kubenswrapper[4814]: E0227 17:20:13.488377 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:20:26 crc kubenswrapper[4814]: I0227 17:20:26.487507 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:20:26 crc kubenswrapper[4814]: E0227 17:20:26.488285 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:20:38 crc kubenswrapper[4814]: I0227 17:20:38.497963 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:20:38 crc kubenswrapper[4814]: E0227 17:20:38.499245 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:20:49 crc kubenswrapper[4814]: I0227 17:20:49.488492 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:20:49 crc kubenswrapper[4814]: E0227 17:20:49.489576 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:20:58 crc kubenswrapper[4814]: I0227 17:20:58.507231 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dnhdp"] Feb 27 17:20:58 crc kubenswrapper[4814]: E0227 17:20:58.508165 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42a61257-51ec-49cd-9e0f-3f872cc8a93b" containerName="oc" Feb 27 17:20:58 crc kubenswrapper[4814]: I0227 17:20:58.508178 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="42a61257-51ec-49cd-9e0f-3f872cc8a93b" containerName="oc" Feb 27 17:20:58 crc kubenswrapper[4814]: I0227 17:20:58.508524 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="42a61257-51ec-49cd-9e0f-3f872cc8a93b" containerName="oc" Feb 27 17:20:58 crc kubenswrapper[4814]: I0227 17:20:58.510676 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:20:58 crc kubenswrapper[4814]: I0227 17:20:58.515023 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dnhdp"] Feb 27 17:20:58 crc kubenswrapper[4814]: I0227 17:20:58.554324 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-catalog-content\") pod \"redhat-operators-dnhdp\" (UID: \"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b\") " pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:20:58 crc kubenswrapper[4814]: I0227 17:20:58.556356 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxx6n\" (UniqueName: \"kubernetes.io/projected/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-kube-api-access-dxx6n\") pod \"redhat-operators-dnhdp\" (UID: \"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b\") " pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:20:58 crc kubenswrapper[4814]: I0227 17:20:58.556755 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-utilities\") pod \"redhat-operators-dnhdp\" (UID: \"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b\") " pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:20:58 crc kubenswrapper[4814]: I0227 17:20:58.659973 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-catalog-content\") pod \"redhat-operators-dnhdp\" (UID: \"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b\") " pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:20:58 crc kubenswrapper[4814]: I0227 17:20:58.660177 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxx6n\" (UniqueName: \"kubernetes.io/projected/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-kube-api-access-dxx6n\") pod \"redhat-operators-dnhdp\" (UID: \"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b\") " pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:20:58 crc kubenswrapper[4814]: I0227 17:20:58.660252 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-utilities\") pod \"redhat-operators-dnhdp\" (UID: \"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b\") " pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:20:58 crc kubenswrapper[4814]: I0227 17:20:58.660815 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-catalog-content\") pod \"redhat-operators-dnhdp\" (UID: \"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b\") " pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:20:58 crc kubenswrapper[4814]: I0227 17:20:58.660860 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-utilities\") pod \"redhat-operators-dnhdp\" (UID: \"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b\") " pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:20:58 crc kubenswrapper[4814]: I0227 17:20:58.687546 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxx6n\" (UniqueName: \"kubernetes.io/projected/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-kube-api-access-dxx6n\") pod \"redhat-operators-dnhdp\" (UID: \"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b\") " pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:20:58 crc kubenswrapper[4814]: I0227 17:20:58.854908 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:20:59 crc kubenswrapper[4814]: I0227 17:20:59.395205 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dnhdp"] Feb 27 17:20:59 crc kubenswrapper[4814]: I0227 17:20:59.962483 4814 generic.go:334] "Generic (PLEG): container finished" podID="305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" containerID="b2188746488ff80f6055d69b3250c21c14672c2b930ccdcbb9b05ce84e5f086f" exitCode=0 Feb 27 17:20:59 crc kubenswrapper[4814]: I0227 17:20:59.962551 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnhdp" event={"ID":"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b","Type":"ContainerDied","Data":"b2188746488ff80f6055d69b3250c21c14672c2b930ccdcbb9b05ce84e5f086f"} Feb 27 17:20:59 crc kubenswrapper[4814]: I0227 17:20:59.962624 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnhdp" event={"ID":"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b","Type":"ContainerStarted","Data":"2c10fed246323c758cc480ab01f4102faf9c9f0746167fc9ca2e1617519a0c46"} Feb 27 17:20:59 crc kubenswrapper[4814]: I0227 17:20:59.967049 4814 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:21:01 crc kubenswrapper[4814]: I0227 17:21:01.988199 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnhdp" event={"ID":"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b","Type":"ContainerStarted","Data":"26db8ca0c9c74f1a97f9adf59db2778cba60445edee194e6f38de081493b80b3"} Feb 27 17:21:04 crc kubenswrapper[4814]: I0227 17:21:04.017807 4814 generic.go:334] "Generic (PLEG): container finished" podID="305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" containerID="26db8ca0c9c74f1a97f9adf59db2778cba60445edee194e6f38de081493b80b3" exitCode=0 Feb 27 17:21:04 crc kubenswrapper[4814]: I0227 17:21:04.017855 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnhdp" event={"ID":"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b","Type":"ContainerDied","Data":"26db8ca0c9c74f1a97f9adf59db2778cba60445edee194e6f38de081493b80b3"} Feb 27 17:21:04 crc kubenswrapper[4814]: I0227 17:21:04.488090 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:21:04 crc kubenswrapper[4814]: E0227 17:21:04.488631 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:21:05 crc kubenswrapper[4814]: I0227 17:21:05.028473 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnhdp" event={"ID":"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b","Type":"ContainerStarted","Data":"be2141f4e98ced90fa01304955a55cb863ad8ef9d04d156738d19a47947e3220"} Feb 27 17:21:05 crc kubenswrapper[4814]: I0227 17:21:05.053541 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dnhdp" podStartSLOduration=2.523924814 podStartE2EDuration="7.053521951s" podCreationTimestamp="2026-02-27 17:20:58 +0000 UTC" firstStartedPulling="2026-02-27 17:20:59.966715851 +0000 UTC m=+3472.419340691" lastFinishedPulling="2026-02-27 17:21:04.496312998 +0000 UTC m=+3476.948937828" observedRunningTime="2026-02-27 17:21:05.046584 +0000 UTC m=+3477.499208830" watchObservedRunningTime="2026-02-27 17:21:05.053521951 +0000 UTC m=+3477.506146781" Feb 27 17:21:08 crc kubenswrapper[4814]: I0227 17:21:08.855769 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:21:08 crc kubenswrapper[4814]: I0227 17:21:08.856591 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:21:09 crc kubenswrapper[4814]: I0227 17:21:09.911970 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dnhdp" podUID="305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" containerName="registry-server" probeResult="failure" output=< Feb 27 17:21:09 crc kubenswrapper[4814]: timeout: failed to connect service ":50051" within 1s Feb 27 17:21:09 crc kubenswrapper[4814]: > Feb 27 17:21:16 crc kubenswrapper[4814]: I0227 17:21:16.487760 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:21:16 crc kubenswrapper[4814]: E0227 17:21:16.488495 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:21:19 crc kubenswrapper[4814]: I0227 17:21:19.908235 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dnhdp" podUID="305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" containerName="registry-server" probeResult="failure" output=< Feb 27 17:21:19 crc kubenswrapper[4814]: timeout: failed to connect service ":50051" within 1s Feb 27 17:21:19 crc kubenswrapper[4814]: > Feb 27 17:21:27 crc kubenswrapper[4814]: I0227 17:21:27.487561 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:21:27 crc kubenswrapper[4814]: E0227 17:21:27.488612 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:21:28 crc kubenswrapper[4814]: I0227 17:21:28.932240 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:21:29 crc kubenswrapper[4814]: I0227 17:21:29.006673 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:21:29 crc kubenswrapper[4814]: I0227 17:21:29.690919 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dnhdp"] Feb 27 17:21:30 crc kubenswrapper[4814]: I0227 17:21:30.280924 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dnhdp" podUID="305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" containerName="registry-server" containerID="cri-o://be2141f4e98ced90fa01304955a55cb863ad8ef9d04d156738d19a47947e3220" gracePeriod=2 Feb 27 17:21:30 crc kubenswrapper[4814]: I0227 17:21:30.794271 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:21:30 crc kubenswrapper[4814]: I0227 17:21:30.985973 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxx6n\" (UniqueName: \"kubernetes.io/projected/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-kube-api-access-dxx6n\") pod \"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b\" (UID: \"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b\") " Feb 27 17:21:30 crc kubenswrapper[4814]: I0227 17:21:30.986093 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-catalog-content\") pod \"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b\" (UID: \"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b\") " Feb 27 17:21:30 crc kubenswrapper[4814]: I0227 17:21:30.986240 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-utilities\") pod \"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b\" (UID: \"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b\") " Feb 27 17:21:30 crc kubenswrapper[4814]: I0227 17:21:30.987493 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-utilities" (OuterVolumeSpecName: "utilities") pod "305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" (UID: "305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:21:30 crc kubenswrapper[4814]: I0227 17:21:30.994523 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-kube-api-access-dxx6n" (OuterVolumeSpecName: "kube-api-access-dxx6n") pod "305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" (UID: "305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b"). InnerVolumeSpecName "kube-api-access-dxx6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.089386 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxx6n\" (UniqueName: \"kubernetes.io/projected/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-kube-api-access-dxx6n\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.089427 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.105219 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" (UID: "305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.191581 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.292355 4814 generic.go:334] "Generic (PLEG): container finished" podID="305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" containerID="be2141f4e98ced90fa01304955a55cb863ad8ef9d04d156738d19a47947e3220" exitCode=0 Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.292406 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnhdp" event={"ID":"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b","Type":"ContainerDied","Data":"be2141f4e98ced90fa01304955a55cb863ad8ef9d04d156738d19a47947e3220"} Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.292452 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnhdp" event={"ID":"305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b","Type":"ContainerDied","Data":"2c10fed246323c758cc480ab01f4102faf9c9f0746167fc9ca2e1617519a0c46"} Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.292486 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dnhdp" Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.292499 4814 scope.go:117] "RemoveContainer" containerID="be2141f4e98ced90fa01304955a55cb863ad8ef9d04d156738d19a47947e3220" Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.331975 4814 scope.go:117] "RemoveContainer" containerID="26db8ca0c9c74f1a97f9adf59db2778cba60445edee194e6f38de081493b80b3" Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.359758 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dnhdp"] Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.369747 4814 scope.go:117] "RemoveContainer" containerID="b2188746488ff80f6055d69b3250c21c14672c2b930ccdcbb9b05ce84e5f086f" Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.373684 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dnhdp"] Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.416426 4814 scope.go:117] "RemoveContainer" containerID="be2141f4e98ced90fa01304955a55cb863ad8ef9d04d156738d19a47947e3220" Feb 27 17:21:31 crc kubenswrapper[4814]: E0227 17:21:31.417030 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be2141f4e98ced90fa01304955a55cb863ad8ef9d04d156738d19a47947e3220\": container with ID starting with be2141f4e98ced90fa01304955a55cb863ad8ef9d04d156738d19a47947e3220 not found: ID does not exist" containerID="be2141f4e98ced90fa01304955a55cb863ad8ef9d04d156738d19a47947e3220" Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.417083 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be2141f4e98ced90fa01304955a55cb863ad8ef9d04d156738d19a47947e3220"} err="failed to get container status \"be2141f4e98ced90fa01304955a55cb863ad8ef9d04d156738d19a47947e3220\": rpc error: code = NotFound desc = could not find container \"be2141f4e98ced90fa01304955a55cb863ad8ef9d04d156738d19a47947e3220\": container with ID starting with be2141f4e98ced90fa01304955a55cb863ad8ef9d04d156738d19a47947e3220 not found: ID does not exist" Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.417117 4814 scope.go:117] "RemoveContainer" containerID="26db8ca0c9c74f1a97f9adf59db2778cba60445edee194e6f38de081493b80b3" Feb 27 17:21:31 crc kubenswrapper[4814]: E0227 17:21:31.417686 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26db8ca0c9c74f1a97f9adf59db2778cba60445edee194e6f38de081493b80b3\": container with ID starting with 26db8ca0c9c74f1a97f9adf59db2778cba60445edee194e6f38de081493b80b3 not found: ID does not exist" containerID="26db8ca0c9c74f1a97f9adf59db2778cba60445edee194e6f38de081493b80b3" Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.417719 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26db8ca0c9c74f1a97f9adf59db2778cba60445edee194e6f38de081493b80b3"} err="failed to get container status \"26db8ca0c9c74f1a97f9adf59db2778cba60445edee194e6f38de081493b80b3\": rpc error: code = NotFound desc = could not find container \"26db8ca0c9c74f1a97f9adf59db2778cba60445edee194e6f38de081493b80b3\": container with ID starting with 26db8ca0c9c74f1a97f9adf59db2778cba60445edee194e6f38de081493b80b3 not found: ID does not exist" Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.417759 4814 scope.go:117] "RemoveContainer" containerID="b2188746488ff80f6055d69b3250c21c14672c2b930ccdcbb9b05ce84e5f086f" Feb 27 17:21:31 crc kubenswrapper[4814]: E0227 17:21:31.418282 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2188746488ff80f6055d69b3250c21c14672c2b930ccdcbb9b05ce84e5f086f\": container with ID starting with b2188746488ff80f6055d69b3250c21c14672c2b930ccdcbb9b05ce84e5f086f not found: ID does not exist" containerID="b2188746488ff80f6055d69b3250c21c14672c2b930ccdcbb9b05ce84e5f086f" Feb 27 17:21:31 crc kubenswrapper[4814]: I0227 17:21:31.418314 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2188746488ff80f6055d69b3250c21c14672c2b930ccdcbb9b05ce84e5f086f"} err="failed to get container status \"b2188746488ff80f6055d69b3250c21c14672c2b930ccdcbb9b05ce84e5f086f\": rpc error: code = NotFound desc = could not find container \"b2188746488ff80f6055d69b3250c21c14672c2b930ccdcbb9b05ce84e5f086f\": container with ID starting with b2188746488ff80f6055d69b3250c21c14672c2b930ccdcbb9b05ce84e5f086f not found: ID does not exist" Feb 27 17:21:32 crc kubenswrapper[4814]: I0227 17:21:32.503515 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" path="/var/lib/kubelet/pods/305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b/volumes" Feb 27 17:21:41 crc kubenswrapper[4814]: I0227 17:21:41.488133 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:21:41 crc kubenswrapper[4814]: E0227 17:21:41.489432 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:21:56 crc kubenswrapper[4814]: I0227 17:21:56.489767 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:21:57 crc kubenswrapper[4814]: I0227 17:21:57.656283 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"daa4ecfa76d33456eca88bbfe3cf3adb033c01a242837064c2ae775d74b120a7"} Feb 27 17:22:00 crc kubenswrapper[4814]: I0227 17:22:00.166975 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536882-pj72p"] Feb 27 17:22:00 crc kubenswrapper[4814]: E0227 17:22:00.169468 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" containerName="extract-content" Feb 27 17:22:00 crc kubenswrapper[4814]: I0227 17:22:00.169513 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" containerName="extract-content" Feb 27 17:22:00 crc kubenswrapper[4814]: E0227 17:22:00.169543 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" containerName="registry-server" Feb 27 17:22:00 crc kubenswrapper[4814]: I0227 17:22:00.169566 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" containerName="registry-server" Feb 27 17:22:00 crc kubenswrapper[4814]: E0227 17:22:00.169644 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" containerName="extract-utilities" Feb 27 17:22:00 crc kubenswrapper[4814]: I0227 17:22:00.169659 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" containerName="extract-utilities" Feb 27 17:22:00 crc kubenswrapper[4814]: I0227 17:22:00.170085 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="305f2a58-eee3-4fc1-87c6-3ad1b8e0d33b" containerName="registry-server" Feb 27 17:22:00 crc kubenswrapper[4814]: I0227 17:22:00.171579 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536882-pj72p" Feb 27 17:22:00 crc kubenswrapper[4814]: I0227 17:22:00.175423 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:22:00 crc kubenswrapper[4814]: I0227 17:22:00.175549 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:22:00 crc kubenswrapper[4814]: I0227 17:22:00.176546 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:22:00 crc kubenswrapper[4814]: I0227 17:22:00.179216 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536882-pj72p"] Feb 27 17:22:00 crc kubenswrapper[4814]: I0227 17:22:00.273469 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szrn2\" (UniqueName: \"kubernetes.io/projected/16bef484-7466-490c-80ec-55001c91d531-kube-api-access-szrn2\") pod \"auto-csr-approver-29536882-pj72p\" (UID: \"16bef484-7466-490c-80ec-55001c91d531\") " pod="openshift-infra/auto-csr-approver-29536882-pj72p" Feb 27 17:22:00 crc kubenswrapper[4814]: I0227 17:22:00.376791 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szrn2\" (UniqueName: \"kubernetes.io/projected/16bef484-7466-490c-80ec-55001c91d531-kube-api-access-szrn2\") pod \"auto-csr-approver-29536882-pj72p\" (UID: \"16bef484-7466-490c-80ec-55001c91d531\") " pod="openshift-infra/auto-csr-approver-29536882-pj72p" Feb 27 17:22:00 crc kubenswrapper[4814]: I0227 17:22:00.402675 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szrn2\" (UniqueName: \"kubernetes.io/projected/16bef484-7466-490c-80ec-55001c91d531-kube-api-access-szrn2\") pod \"auto-csr-approver-29536882-pj72p\" (UID: \"16bef484-7466-490c-80ec-55001c91d531\") " pod="openshift-infra/auto-csr-approver-29536882-pj72p" Feb 27 17:22:00 crc kubenswrapper[4814]: I0227 17:22:00.512784 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536882-pj72p" Feb 27 17:22:01 crc kubenswrapper[4814]: I0227 17:22:01.056037 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536882-pj72p"] Feb 27 17:22:01 crc kubenswrapper[4814]: W0227 17:22:01.062697 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16bef484_7466_490c_80ec_55001c91d531.slice/crio-f0ae8e4346de8b6e2f41ee71ea091816cd52b88c600c8c857af797f6002b0faf WatchSource:0}: Error finding container f0ae8e4346de8b6e2f41ee71ea091816cd52b88c600c8c857af797f6002b0faf: Status 404 returned error can't find the container with id f0ae8e4346de8b6e2f41ee71ea091816cd52b88c600c8c857af797f6002b0faf Feb 27 17:22:01 crc kubenswrapper[4814]: I0227 17:22:01.710890 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536882-pj72p" event={"ID":"16bef484-7466-490c-80ec-55001c91d531","Type":"ContainerStarted","Data":"f0ae8e4346de8b6e2f41ee71ea091816cd52b88c600c8c857af797f6002b0faf"} Feb 27 17:22:02 crc kubenswrapper[4814]: I0227 17:22:02.723674 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536882-pj72p" event={"ID":"16bef484-7466-490c-80ec-55001c91d531","Type":"ContainerStarted","Data":"d2708a39c6e6743ab5bad4a16e91735a1af99b89f2e978a63babb49df312533d"} Feb 27 17:22:02 crc kubenswrapper[4814]: I0227 17:22:02.743655 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536882-pj72p" podStartSLOduration=1.590664942 podStartE2EDuration="2.743626864s" podCreationTimestamp="2026-02-27 17:22:00 +0000 UTC" firstStartedPulling="2026-02-27 17:22:01.066111747 +0000 UTC m=+3533.518736587" lastFinishedPulling="2026-02-27 17:22:02.219073659 +0000 UTC m=+3534.671698509" observedRunningTime="2026-02-27 17:22:02.737775645 +0000 UTC m=+3535.190400505" watchObservedRunningTime="2026-02-27 17:22:02.743626864 +0000 UTC m=+3535.196251714" Feb 27 17:22:03 crc kubenswrapper[4814]: I0227 17:22:03.733816 4814 generic.go:334] "Generic (PLEG): container finished" podID="16bef484-7466-490c-80ec-55001c91d531" containerID="d2708a39c6e6743ab5bad4a16e91735a1af99b89f2e978a63babb49df312533d" exitCode=0 Feb 27 17:22:03 crc kubenswrapper[4814]: I0227 17:22:03.733863 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536882-pj72p" event={"ID":"16bef484-7466-490c-80ec-55001c91d531","Type":"ContainerDied","Data":"d2708a39c6e6743ab5bad4a16e91735a1af99b89f2e978a63babb49df312533d"} Feb 27 17:22:05 crc kubenswrapper[4814]: I0227 17:22:05.202904 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536882-pj72p" Feb 27 17:22:05 crc kubenswrapper[4814]: I0227 17:22:05.300161 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szrn2\" (UniqueName: \"kubernetes.io/projected/16bef484-7466-490c-80ec-55001c91d531-kube-api-access-szrn2\") pod \"16bef484-7466-490c-80ec-55001c91d531\" (UID: \"16bef484-7466-490c-80ec-55001c91d531\") " Feb 27 17:22:05 crc kubenswrapper[4814]: I0227 17:22:05.310399 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16bef484-7466-490c-80ec-55001c91d531-kube-api-access-szrn2" (OuterVolumeSpecName: "kube-api-access-szrn2") pod "16bef484-7466-490c-80ec-55001c91d531" (UID: "16bef484-7466-490c-80ec-55001c91d531"). InnerVolumeSpecName "kube-api-access-szrn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:05 crc kubenswrapper[4814]: I0227 17:22:05.403689 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szrn2\" (UniqueName: \"kubernetes.io/projected/16bef484-7466-490c-80ec-55001c91d531-kube-api-access-szrn2\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:05 crc kubenswrapper[4814]: I0227 17:22:05.769686 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536882-pj72p" event={"ID":"16bef484-7466-490c-80ec-55001c91d531","Type":"ContainerDied","Data":"f0ae8e4346de8b6e2f41ee71ea091816cd52b88c600c8c857af797f6002b0faf"} Feb 27 17:22:05 crc kubenswrapper[4814]: I0227 17:22:05.769758 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0ae8e4346de8b6e2f41ee71ea091816cd52b88c600c8c857af797f6002b0faf" Feb 27 17:22:05 crc kubenswrapper[4814]: I0227 17:22:05.769837 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536882-pj72p" Feb 27 17:22:05 crc kubenswrapper[4814]: I0227 17:22:05.844646 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536876-c6xqw"] Feb 27 17:22:05 crc kubenswrapper[4814]: I0227 17:22:05.855805 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536876-c6xqw"] Feb 27 17:22:06 crc kubenswrapper[4814]: I0227 17:22:06.507063 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bfb3c95-d26c-4660-a3f7-208d65f53ea1" path="/var/lib/kubelet/pods/8bfb3c95-d26c-4660-a3f7-208d65f53ea1/volumes" Feb 27 17:22:07 crc kubenswrapper[4814]: I0227 17:22:07.128952 4814 scope.go:117] "RemoveContainer" containerID="2e168673b962f2c6971291c455865f9e21b4b8800a761d0e3927a94a5a5a66c9" Feb 27 17:23:27 crc kubenswrapper[4814]: I0227 17:23:27.902322 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kbbd8"] Feb 27 17:23:27 crc kubenswrapper[4814]: E0227 17:23:27.903833 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16bef484-7466-490c-80ec-55001c91d531" containerName="oc" Feb 27 17:23:27 crc kubenswrapper[4814]: I0227 17:23:27.903856 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="16bef484-7466-490c-80ec-55001c91d531" containerName="oc" Feb 27 17:23:27 crc kubenswrapper[4814]: I0227 17:23:27.904129 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="16bef484-7466-490c-80ec-55001c91d531" containerName="oc" Feb 27 17:23:27 crc kubenswrapper[4814]: I0227 17:23:27.905758 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbbd8" Feb 27 17:23:27 crc kubenswrapper[4814]: I0227 17:23:27.928568 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kbbd8"] Feb 27 17:23:27 crc kubenswrapper[4814]: I0227 17:23:27.993095 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/705d8544-aed0-40dd-aac2-185cfc8ecb06-catalog-content\") pod \"certified-operators-kbbd8\" (UID: \"705d8544-aed0-40dd-aac2-185cfc8ecb06\") " pod="openshift-marketplace/certified-operators-kbbd8" Feb 27 17:23:27 crc kubenswrapper[4814]: I0227 17:23:27.993170 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrdss\" (UniqueName: \"kubernetes.io/projected/705d8544-aed0-40dd-aac2-185cfc8ecb06-kube-api-access-hrdss\") pod \"certified-operators-kbbd8\" (UID: \"705d8544-aed0-40dd-aac2-185cfc8ecb06\") " pod="openshift-marketplace/certified-operators-kbbd8" Feb 27 17:23:27 crc kubenswrapper[4814]: I0227 17:23:27.993306 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/705d8544-aed0-40dd-aac2-185cfc8ecb06-utilities\") pod \"certified-operators-kbbd8\" (UID: \"705d8544-aed0-40dd-aac2-185cfc8ecb06\") " pod="openshift-marketplace/certified-operators-kbbd8" Feb 27 17:23:28 crc kubenswrapper[4814]: I0227 17:23:28.095564 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/705d8544-aed0-40dd-aac2-185cfc8ecb06-utilities\") pod \"certified-operators-kbbd8\" (UID: \"705d8544-aed0-40dd-aac2-185cfc8ecb06\") " pod="openshift-marketplace/certified-operators-kbbd8" Feb 27 17:23:28 crc kubenswrapper[4814]: I0227 17:23:28.095693 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/705d8544-aed0-40dd-aac2-185cfc8ecb06-catalog-content\") pod \"certified-operators-kbbd8\" (UID: \"705d8544-aed0-40dd-aac2-185cfc8ecb06\") " pod="openshift-marketplace/certified-operators-kbbd8" Feb 27 17:23:28 crc kubenswrapper[4814]: I0227 17:23:28.095742 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrdss\" (UniqueName: \"kubernetes.io/projected/705d8544-aed0-40dd-aac2-185cfc8ecb06-kube-api-access-hrdss\") pod \"certified-operators-kbbd8\" (UID: \"705d8544-aed0-40dd-aac2-185cfc8ecb06\") " pod="openshift-marketplace/certified-operators-kbbd8" Feb 27 17:23:28 crc kubenswrapper[4814]: I0227 17:23:28.096614 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/705d8544-aed0-40dd-aac2-185cfc8ecb06-utilities\") pod \"certified-operators-kbbd8\" (UID: \"705d8544-aed0-40dd-aac2-185cfc8ecb06\") " pod="openshift-marketplace/certified-operators-kbbd8" Feb 27 17:23:28 crc kubenswrapper[4814]: I0227 17:23:28.096828 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/705d8544-aed0-40dd-aac2-185cfc8ecb06-catalog-content\") pod \"certified-operators-kbbd8\" (UID: \"705d8544-aed0-40dd-aac2-185cfc8ecb06\") " pod="openshift-marketplace/certified-operators-kbbd8" Feb 27 17:23:28 crc kubenswrapper[4814]: I0227 17:23:28.121018 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrdss\" (UniqueName: \"kubernetes.io/projected/705d8544-aed0-40dd-aac2-185cfc8ecb06-kube-api-access-hrdss\") pod \"certified-operators-kbbd8\" (UID: \"705d8544-aed0-40dd-aac2-185cfc8ecb06\") " pod="openshift-marketplace/certified-operators-kbbd8" Feb 27 17:23:28 crc kubenswrapper[4814]: I0227 17:23:28.248465 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbbd8" Feb 27 17:23:28 crc kubenswrapper[4814]: I0227 17:23:28.737837 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kbbd8"] Feb 27 17:23:28 crc kubenswrapper[4814]: I0227 17:23:28.800927 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbbd8" event={"ID":"705d8544-aed0-40dd-aac2-185cfc8ecb06","Type":"ContainerStarted","Data":"16f6ffdaa5e0a6b541464c00ca8a5598e6844c13a718c269fa9b192e2d4c4cff"} Feb 27 17:23:29 crc kubenswrapper[4814]: I0227 17:23:29.824900 4814 generic.go:334] "Generic (PLEG): container finished" podID="705d8544-aed0-40dd-aac2-185cfc8ecb06" containerID="b13a2698e2c77af7ae1d886fda65fa701a760d42a5a8b278a2addf9f758cc743" exitCode=0 Feb 27 17:23:29 crc kubenswrapper[4814]: I0227 17:23:29.825792 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbbd8" event={"ID":"705d8544-aed0-40dd-aac2-185cfc8ecb06","Type":"ContainerDied","Data":"b13a2698e2c77af7ae1d886fda65fa701a760d42a5a8b278a2addf9f758cc743"} Feb 27 17:23:35 crc kubenswrapper[4814]: I0227 17:23:35.933388 4814 generic.go:334] "Generic (PLEG): container finished" podID="705d8544-aed0-40dd-aac2-185cfc8ecb06" containerID="29afedede7cc30dfcc322fd239b86c9d92bc9020a268d389c49a0b658e856180" exitCode=0 Feb 27 17:23:35 crc kubenswrapper[4814]: I0227 17:23:35.933478 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbbd8" event={"ID":"705d8544-aed0-40dd-aac2-185cfc8ecb06","Type":"ContainerDied","Data":"29afedede7cc30dfcc322fd239b86c9d92bc9020a268d389c49a0b658e856180"} Feb 27 17:23:35 crc kubenswrapper[4814]: I0227 17:23:35.937015 4814 generic.go:334] "Generic (PLEG): container finished" podID="c11159aa-3255-4574-81a3-182c8a30d970" containerID="f837f3034c30072938725c0f2ba43c8dce3eb86064b075d33ec72f1e30337bcc" exitCode=0 Feb 27 17:23:35 crc kubenswrapper[4814]: I0227 17:23:35.937051 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c11159aa-3255-4574-81a3-182c8a30d970","Type":"ContainerDied","Data":"f837f3034c30072938725c0f2ba43c8dce3eb86064b075d33ec72f1e30337bcc"} Feb 27 17:23:36 crc kubenswrapper[4814]: I0227 17:23:36.957159 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbbd8" event={"ID":"705d8544-aed0-40dd-aac2-185cfc8ecb06","Type":"ContainerStarted","Data":"fa402fab40a8694522040e0e618245f6d661bf04c72858a32970ee556f4dc463"} Feb 27 17:23:36 crc kubenswrapper[4814]: I0227 17:23:36.985727 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kbbd8" podStartSLOduration=3.392232227 podStartE2EDuration="9.985694954s" podCreationTimestamp="2026-02-27 17:23:27 +0000 UTC" firstStartedPulling="2026-02-27 17:23:29.829715022 +0000 UTC m=+3622.282339872" lastFinishedPulling="2026-02-27 17:23:36.423177769 +0000 UTC m=+3628.875802599" observedRunningTime="2026-02-27 17:23:36.980190426 +0000 UTC m=+3629.432815266" watchObservedRunningTime="2026-02-27 17:23:36.985694954 +0000 UTC m=+3629.438319814" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.479372 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.657828 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-ssh-key\") pod \"c11159aa-3255-4574-81a3-182c8a30d970\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.657876 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-ca-certs\") pod \"c11159aa-3255-4574-81a3-182c8a30d970\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.657952 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-openstack-config-secret\") pod \"c11159aa-3255-4574-81a3-182c8a30d970\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.657986 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9r85q\" (UniqueName: \"kubernetes.io/projected/c11159aa-3255-4574-81a3-182c8a30d970-kube-api-access-9r85q\") pod \"c11159aa-3255-4574-81a3-182c8a30d970\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.658052 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c11159aa-3255-4574-81a3-182c8a30d970-test-operator-ephemeral-temporary\") pod \"c11159aa-3255-4574-81a3-182c8a30d970\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.658092 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c11159aa-3255-4574-81a3-182c8a30d970-config-data\") pod \"c11159aa-3255-4574-81a3-182c8a30d970\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.658210 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c11159aa-3255-4574-81a3-182c8a30d970-openstack-config\") pod \"c11159aa-3255-4574-81a3-182c8a30d970\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.658237 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"c11159aa-3255-4574-81a3-182c8a30d970\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.658356 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c11159aa-3255-4574-81a3-182c8a30d970-test-operator-ephemeral-workdir\") pod \"c11159aa-3255-4574-81a3-182c8a30d970\" (UID: \"c11159aa-3255-4574-81a3-182c8a30d970\") " Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.659332 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c11159aa-3255-4574-81a3-182c8a30d970-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "c11159aa-3255-4574-81a3-182c8a30d970" (UID: "c11159aa-3255-4574-81a3-182c8a30d970"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.659396 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c11159aa-3255-4574-81a3-182c8a30d970-config-data" (OuterVolumeSpecName: "config-data") pod "c11159aa-3255-4574-81a3-182c8a30d970" (UID: "c11159aa-3255-4574-81a3-182c8a30d970"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.664992 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c11159aa-3255-4574-81a3-182c8a30d970-kube-api-access-9r85q" (OuterVolumeSpecName: "kube-api-access-9r85q") pod "c11159aa-3255-4574-81a3-182c8a30d970" (UID: "c11159aa-3255-4574-81a3-182c8a30d970"). InnerVolumeSpecName "kube-api-access-9r85q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.668046 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "test-operator-logs") pod "c11159aa-3255-4574-81a3-182c8a30d970" (UID: "c11159aa-3255-4574-81a3-182c8a30d970"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.669502 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c11159aa-3255-4574-81a3-182c8a30d970-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "c11159aa-3255-4574-81a3-182c8a30d970" (UID: "c11159aa-3255-4574-81a3-182c8a30d970"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.692774 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "c11159aa-3255-4574-81a3-182c8a30d970" (UID: "c11159aa-3255-4574-81a3-182c8a30d970"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.693288 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "c11159aa-3255-4574-81a3-182c8a30d970" (UID: "c11159aa-3255-4574-81a3-182c8a30d970"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.711038 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c11159aa-3255-4574-81a3-182c8a30d970" (UID: "c11159aa-3255-4574-81a3-182c8a30d970"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.731750 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c11159aa-3255-4574-81a3-182c8a30d970-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "c11159aa-3255-4574-81a3-182c8a30d970" (UID: "c11159aa-3255-4574-81a3-182c8a30d970"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.761277 4814 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-ssh-key\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.761317 4814 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-ca-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.761328 4814 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c11159aa-3255-4574-81a3-182c8a30d970-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.761343 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9r85q\" (UniqueName: \"kubernetes.io/projected/c11159aa-3255-4574-81a3-182c8a30d970-kube-api-access-9r85q\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.761355 4814 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c11159aa-3255-4574-81a3-182c8a30d970-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.761366 4814 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c11159aa-3255-4574-81a3-182c8a30d970-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.761377 4814 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c11159aa-3255-4574-81a3-182c8a30d970-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.761418 4814 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.761438 4814 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c11159aa-3255-4574-81a3-182c8a30d970-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.791319 4814 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.864474 4814 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.968902 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c11159aa-3255-4574-81a3-182c8a30d970","Type":"ContainerDied","Data":"fb542b3688380e6d3c1742f0ae5c7794c81dc270fe03e7866576ac197c341bd9"} Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.968958 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb542b3688380e6d3c1742f0ae5c7794c81dc270fe03e7866576ac197c341bd9" Feb 27 17:23:37 crc kubenswrapper[4814]: I0227 17:23:37.968981 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 27 17:23:38 crc kubenswrapper[4814]: I0227 17:23:38.249596 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kbbd8" Feb 27 17:23:38 crc kubenswrapper[4814]: I0227 17:23:38.249666 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kbbd8" Feb 27 17:23:39 crc kubenswrapper[4814]: I0227 17:23:39.307764 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-kbbd8" podUID="705d8544-aed0-40dd-aac2-185cfc8ecb06" containerName="registry-server" probeResult="failure" output=< Feb 27 17:23:39 crc kubenswrapper[4814]: timeout: failed to connect service ":50051" within 1s Feb 27 17:23:39 crc kubenswrapper[4814]: > Feb 27 17:23:47 crc kubenswrapper[4814]: I0227 17:23:47.287077 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 27 17:23:47 crc kubenswrapper[4814]: E0227 17:23:47.288483 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c11159aa-3255-4574-81a3-182c8a30d970" containerName="tempest-tests-tempest-tests-runner" Feb 27 17:23:47 crc kubenswrapper[4814]: I0227 17:23:47.288505 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="c11159aa-3255-4574-81a3-182c8a30d970" containerName="tempest-tests-tempest-tests-runner" Feb 27 17:23:47 crc kubenswrapper[4814]: I0227 17:23:47.288751 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="c11159aa-3255-4574-81a3-182c8a30d970" containerName="tempest-tests-tempest-tests-runner" Feb 27 17:23:47 crc kubenswrapper[4814]: I0227 17:23:47.289692 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 17:23:47 crc kubenswrapper[4814]: I0227 17:23:47.293465 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cvqcf" Feb 27 17:23:47 crc kubenswrapper[4814]: I0227 17:23:47.306988 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 27 17:23:47 crc kubenswrapper[4814]: I0227 17:23:47.414184 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lk9n\" (UniqueName: \"kubernetes.io/projected/19e7467f-d219-4179-8f27-c3f0ecc2041f-kube-api-access-6lk9n\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"19e7467f-d219-4179-8f27-c3f0ecc2041f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 17:23:47 crc kubenswrapper[4814]: I0227 17:23:47.414730 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"19e7467f-d219-4179-8f27-c3f0ecc2041f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 17:23:47 crc kubenswrapper[4814]: I0227 17:23:47.517203 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lk9n\" (UniqueName: \"kubernetes.io/projected/19e7467f-d219-4179-8f27-c3f0ecc2041f-kube-api-access-6lk9n\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"19e7467f-d219-4179-8f27-c3f0ecc2041f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 17:23:47 crc kubenswrapper[4814]: I0227 17:23:47.517922 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"19e7467f-d219-4179-8f27-c3f0ecc2041f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 17:23:47 crc kubenswrapper[4814]: I0227 17:23:47.518501 4814 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"19e7467f-d219-4179-8f27-c3f0ecc2041f\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 17:23:47 crc kubenswrapper[4814]: I0227 17:23:47.541077 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lk9n\" (UniqueName: \"kubernetes.io/projected/19e7467f-d219-4179-8f27-c3f0ecc2041f-kube-api-access-6lk9n\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"19e7467f-d219-4179-8f27-c3f0ecc2041f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 17:23:47 crc kubenswrapper[4814]: I0227 17:23:47.552551 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"19e7467f-d219-4179-8f27-c3f0ecc2041f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 17:23:47 crc kubenswrapper[4814]: I0227 17:23:47.637152 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 17:23:48 crc kubenswrapper[4814]: I0227 17:23:48.161198 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 27 17:23:48 crc kubenswrapper[4814]: W0227 17:23:48.163931 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19e7467f_d219_4179_8f27_c3f0ecc2041f.slice/crio-bdf07b11de515b54aca2591fe916309db40ba6ec5a92e7ab35127be305fb178f WatchSource:0}: Error finding container bdf07b11de515b54aca2591fe916309db40ba6ec5a92e7ab35127be305fb178f: Status 404 returned error can't find the container with id bdf07b11de515b54aca2591fe916309db40ba6ec5a92e7ab35127be305fb178f Feb 27 17:23:48 crc kubenswrapper[4814]: I0227 17:23:48.308687 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kbbd8" Feb 27 17:23:48 crc kubenswrapper[4814]: I0227 17:23:48.359222 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kbbd8" Feb 27 17:23:48 crc kubenswrapper[4814]: I0227 17:23:48.449881 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kbbd8"] Feb 27 17:23:48 crc kubenswrapper[4814]: I0227 17:23:48.558124 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4nz7s"] Feb 27 17:23:48 crc kubenswrapper[4814]: I0227 17:23:48.558453 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4nz7s" podUID="d7b69f1c-8ca8-472d-9357-0350e8d732d4" containerName="registry-server" containerID="cri-o://e13d476b1802f964dc3a86730c1f816149b7d18cae7d6acb6521cb0c64817f4a" gracePeriod=2 Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.011125 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.111873 4814 generic.go:334] "Generic (PLEG): container finished" podID="d7b69f1c-8ca8-472d-9357-0350e8d732d4" containerID="e13d476b1802f964dc3a86730c1f816149b7d18cae7d6acb6521cb0c64817f4a" exitCode=0 Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.111995 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nz7s" event={"ID":"d7b69f1c-8ca8-472d-9357-0350e8d732d4","Type":"ContainerDied","Data":"e13d476b1802f964dc3a86730c1f816149b7d18cae7d6acb6521cb0c64817f4a"} Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.112033 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nz7s" event={"ID":"d7b69f1c-8ca8-472d-9357-0350e8d732d4","Type":"ContainerDied","Data":"eb96402b588c054e595670e7ac11f855ac8375258207dc2e8d9843a896f08ff6"} Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.112070 4814 scope.go:117] "RemoveContainer" containerID="e13d476b1802f964dc3a86730c1f816149b7d18cae7d6acb6521cb0c64817f4a" Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.112299 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4nz7s" Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.120807 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"19e7467f-d219-4179-8f27-c3f0ecc2041f","Type":"ContainerStarted","Data":"bdf07b11de515b54aca2591fe916309db40ba6ec5a92e7ab35127be305fb178f"} Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.147775 4814 scope.go:117] "RemoveContainer" containerID="cee7dd070e9a8af352e007ab238c67b868a4737799d1e7e446b2b08fe9956b85" Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.159304 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7b69f1c-8ca8-472d-9357-0350e8d732d4-catalog-content\") pod \"d7b69f1c-8ca8-472d-9357-0350e8d732d4\" (UID: \"d7b69f1c-8ca8-472d-9357-0350e8d732d4\") " Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.159762 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5jmh\" (UniqueName: \"kubernetes.io/projected/d7b69f1c-8ca8-472d-9357-0350e8d732d4-kube-api-access-b5jmh\") pod \"d7b69f1c-8ca8-472d-9357-0350e8d732d4\" (UID: \"d7b69f1c-8ca8-472d-9357-0350e8d732d4\") " Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.159853 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7b69f1c-8ca8-472d-9357-0350e8d732d4-utilities\") pod \"d7b69f1c-8ca8-472d-9357-0350e8d732d4\" (UID: \"d7b69f1c-8ca8-472d-9357-0350e8d732d4\") " Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.163074 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7b69f1c-8ca8-472d-9357-0350e8d732d4-utilities" (OuterVolumeSpecName: "utilities") pod "d7b69f1c-8ca8-472d-9357-0350e8d732d4" (UID: "d7b69f1c-8ca8-472d-9357-0350e8d732d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.176476 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7b69f1c-8ca8-472d-9357-0350e8d732d4-kube-api-access-b5jmh" (OuterVolumeSpecName: "kube-api-access-b5jmh") pod "d7b69f1c-8ca8-472d-9357-0350e8d732d4" (UID: "d7b69f1c-8ca8-472d-9357-0350e8d732d4"). InnerVolumeSpecName "kube-api-access-b5jmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.176743 4814 scope.go:117] "RemoveContainer" containerID="e00b746a40de0797ed3a69e7c1a4f0490c4dd159876129fbebba487a5a115ccb" Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.237981 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7b69f1c-8ca8-472d-9357-0350e8d732d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7b69f1c-8ca8-472d-9357-0350e8d732d4" (UID: "d7b69f1c-8ca8-472d-9357-0350e8d732d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.264623 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7b69f1c-8ca8-472d-9357-0350e8d732d4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.264672 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5jmh\" (UniqueName: \"kubernetes.io/projected/d7b69f1c-8ca8-472d-9357-0350e8d732d4-kube-api-access-b5jmh\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.264690 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7b69f1c-8ca8-472d-9357-0350e8d732d4-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.457300 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4nz7s"] Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.465870 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4nz7s"] Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.542603 4814 scope.go:117] "RemoveContainer" containerID="e13d476b1802f964dc3a86730c1f816149b7d18cae7d6acb6521cb0c64817f4a" Feb 27 17:23:49 crc kubenswrapper[4814]: E0227 17:23:49.547655 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e13d476b1802f964dc3a86730c1f816149b7d18cae7d6acb6521cb0c64817f4a\": container with ID starting with e13d476b1802f964dc3a86730c1f816149b7d18cae7d6acb6521cb0c64817f4a not found: ID does not exist" containerID="e13d476b1802f964dc3a86730c1f816149b7d18cae7d6acb6521cb0c64817f4a" Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.547770 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e13d476b1802f964dc3a86730c1f816149b7d18cae7d6acb6521cb0c64817f4a"} err="failed to get container status \"e13d476b1802f964dc3a86730c1f816149b7d18cae7d6acb6521cb0c64817f4a\": rpc error: code = NotFound desc = could not find container \"e13d476b1802f964dc3a86730c1f816149b7d18cae7d6acb6521cb0c64817f4a\": container with ID starting with e13d476b1802f964dc3a86730c1f816149b7d18cae7d6acb6521cb0c64817f4a not found: ID does not exist" Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.547828 4814 scope.go:117] "RemoveContainer" containerID="cee7dd070e9a8af352e007ab238c67b868a4737799d1e7e446b2b08fe9956b85" Feb 27 17:23:49 crc kubenswrapper[4814]: E0227 17:23:49.548248 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cee7dd070e9a8af352e007ab238c67b868a4737799d1e7e446b2b08fe9956b85\": container with ID starting with cee7dd070e9a8af352e007ab238c67b868a4737799d1e7e446b2b08fe9956b85 not found: ID does not exist" containerID="cee7dd070e9a8af352e007ab238c67b868a4737799d1e7e446b2b08fe9956b85" Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.548291 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cee7dd070e9a8af352e007ab238c67b868a4737799d1e7e446b2b08fe9956b85"} err="failed to get container status \"cee7dd070e9a8af352e007ab238c67b868a4737799d1e7e446b2b08fe9956b85\": rpc error: code = NotFound desc = could not find container \"cee7dd070e9a8af352e007ab238c67b868a4737799d1e7e446b2b08fe9956b85\": container with ID starting with cee7dd070e9a8af352e007ab238c67b868a4737799d1e7e446b2b08fe9956b85 not found: ID does not exist" Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.548308 4814 scope.go:117] "RemoveContainer" containerID="e00b746a40de0797ed3a69e7c1a4f0490c4dd159876129fbebba487a5a115ccb" Feb 27 17:23:49 crc kubenswrapper[4814]: E0227 17:23:49.548699 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e00b746a40de0797ed3a69e7c1a4f0490c4dd159876129fbebba487a5a115ccb\": container with ID starting with e00b746a40de0797ed3a69e7c1a4f0490c4dd159876129fbebba487a5a115ccb not found: ID does not exist" containerID="e00b746a40de0797ed3a69e7c1a4f0490c4dd159876129fbebba487a5a115ccb" Feb 27 17:23:49 crc kubenswrapper[4814]: I0227 17:23:49.548722 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e00b746a40de0797ed3a69e7c1a4f0490c4dd159876129fbebba487a5a115ccb"} err="failed to get container status \"e00b746a40de0797ed3a69e7c1a4f0490c4dd159876129fbebba487a5a115ccb\": rpc error: code = NotFound desc = could not find container \"e00b746a40de0797ed3a69e7c1a4f0490c4dd159876129fbebba487a5a115ccb\": container with ID starting with e00b746a40de0797ed3a69e7c1a4f0490c4dd159876129fbebba487a5a115ccb not found: ID does not exist" Feb 27 17:23:50 crc kubenswrapper[4814]: I0227 17:23:50.132127 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"19e7467f-d219-4179-8f27-c3f0ecc2041f","Type":"ContainerStarted","Data":"d4743395a14fee23a120a26190c793291ed20f82d6d72f23ace91acbd0745d97"} Feb 27 17:23:50 crc kubenswrapper[4814]: I0227 17:23:50.152705 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.7316065680000001 podStartE2EDuration="3.152685811s" podCreationTimestamp="2026-02-27 17:23:47 +0000 UTC" firstStartedPulling="2026-02-27 17:23:48.16897334 +0000 UTC m=+3640.621598200" lastFinishedPulling="2026-02-27 17:23:49.590052613 +0000 UTC m=+3642.042677443" observedRunningTime="2026-02-27 17:23:50.147776651 +0000 UTC m=+3642.600401491" watchObservedRunningTime="2026-02-27 17:23:50.152685811 +0000 UTC m=+3642.605310641" Feb 27 17:23:50 crc kubenswrapper[4814]: I0227 17:23:50.498715 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7b69f1c-8ca8-472d-9357-0350e8d732d4" path="/var/lib/kubelet/pods/d7b69f1c-8ca8-472d-9357-0350e8d732d4/volumes" Feb 27 17:24:00 crc kubenswrapper[4814]: I0227 17:24:00.179041 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536884-lb8rx"] Feb 27 17:24:00 crc kubenswrapper[4814]: E0227 17:24:00.180909 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7b69f1c-8ca8-472d-9357-0350e8d732d4" containerName="extract-content" Feb 27 17:24:00 crc kubenswrapper[4814]: I0227 17:24:00.180943 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7b69f1c-8ca8-472d-9357-0350e8d732d4" containerName="extract-content" Feb 27 17:24:00 crc kubenswrapper[4814]: E0227 17:24:00.180995 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7b69f1c-8ca8-472d-9357-0350e8d732d4" containerName="extract-utilities" Feb 27 17:24:00 crc kubenswrapper[4814]: I0227 17:24:00.181009 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7b69f1c-8ca8-472d-9357-0350e8d732d4" containerName="extract-utilities" Feb 27 17:24:00 crc kubenswrapper[4814]: E0227 17:24:00.181042 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7b69f1c-8ca8-472d-9357-0350e8d732d4" containerName="registry-server" Feb 27 17:24:00 crc kubenswrapper[4814]: I0227 17:24:00.181054 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7b69f1c-8ca8-472d-9357-0350e8d732d4" containerName="registry-server" Feb 27 17:24:00 crc kubenswrapper[4814]: I0227 17:24:00.181473 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7b69f1c-8ca8-472d-9357-0350e8d732d4" containerName="registry-server" Feb 27 17:24:00 crc kubenswrapper[4814]: I0227 17:24:00.182728 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536884-lb8rx" Feb 27 17:24:00 crc kubenswrapper[4814]: I0227 17:24:00.186468 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:24:00 crc kubenswrapper[4814]: I0227 17:24:00.188755 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:24:00 crc kubenswrapper[4814]: I0227 17:24:00.189655 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:24:00 crc kubenswrapper[4814]: I0227 17:24:00.196397 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536884-lb8rx"] Feb 27 17:24:00 crc kubenswrapper[4814]: I0227 17:24:00.255795 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfspp\" (UniqueName: \"kubernetes.io/projected/3596843e-9ae6-45eb-b1c9-e82215ced5e6-kube-api-access-pfspp\") pod \"auto-csr-approver-29536884-lb8rx\" (UID: \"3596843e-9ae6-45eb-b1c9-e82215ced5e6\") " pod="openshift-infra/auto-csr-approver-29536884-lb8rx" Feb 27 17:24:00 crc kubenswrapper[4814]: I0227 17:24:00.358110 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfspp\" (UniqueName: \"kubernetes.io/projected/3596843e-9ae6-45eb-b1c9-e82215ced5e6-kube-api-access-pfspp\") pod \"auto-csr-approver-29536884-lb8rx\" (UID: \"3596843e-9ae6-45eb-b1c9-e82215ced5e6\") " pod="openshift-infra/auto-csr-approver-29536884-lb8rx" Feb 27 17:24:00 crc kubenswrapper[4814]: I0227 17:24:00.380901 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfspp\" (UniqueName: \"kubernetes.io/projected/3596843e-9ae6-45eb-b1c9-e82215ced5e6-kube-api-access-pfspp\") pod \"auto-csr-approver-29536884-lb8rx\" (UID: \"3596843e-9ae6-45eb-b1c9-e82215ced5e6\") " pod="openshift-infra/auto-csr-approver-29536884-lb8rx" Feb 27 17:24:00 crc kubenswrapper[4814]: I0227 17:24:00.533747 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536884-lb8rx" Feb 27 17:24:01 crc kubenswrapper[4814]: I0227 17:24:01.053865 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536884-lb8rx"] Feb 27 17:24:01 crc kubenswrapper[4814]: I0227 17:24:01.288958 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536884-lb8rx" event={"ID":"3596843e-9ae6-45eb-b1c9-e82215ced5e6","Type":"ContainerStarted","Data":"499b422985c0d4b559cbf794415df8f2a0d198354a7c4847716efe3386cc494a"} Feb 27 17:24:03 crc kubenswrapper[4814]: I0227 17:24:03.322920 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536884-lb8rx" event={"ID":"3596843e-9ae6-45eb-b1c9-e82215ced5e6","Type":"ContainerStarted","Data":"32160fe46f2873c0b65fc5338bb1f9d69c3d5960d23dcb8b595f7c7112647eaf"} Feb 27 17:24:03 crc kubenswrapper[4814]: I0227 17:24:03.350715 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536884-lb8rx" podStartSLOduration=1.577151175 podStartE2EDuration="3.350688955s" podCreationTimestamp="2026-02-27 17:24:00 +0000 UTC" firstStartedPulling="2026-02-27 17:24:01.069761664 +0000 UTC m=+3653.522386534" lastFinishedPulling="2026-02-27 17:24:02.843299484 +0000 UTC m=+3655.295924314" observedRunningTime="2026-02-27 17:24:03.347619751 +0000 UTC m=+3655.800244591" watchObservedRunningTime="2026-02-27 17:24:03.350688955 +0000 UTC m=+3655.803313795" Feb 27 17:24:04 crc kubenswrapper[4814]: I0227 17:24:04.337858 4814 generic.go:334] "Generic (PLEG): container finished" podID="3596843e-9ae6-45eb-b1c9-e82215ced5e6" containerID="32160fe46f2873c0b65fc5338bb1f9d69c3d5960d23dcb8b595f7c7112647eaf" exitCode=0 Feb 27 17:24:04 crc kubenswrapper[4814]: I0227 17:24:04.338430 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536884-lb8rx" event={"ID":"3596843e-9ae6-45eb-b1c9-e82215ced5e6","Type":"ContainerDied","Data":"32160fe46f2873c0b65fc5338bb1f9d69c3d5960d23dcb8b595f7c7112647eaf"} Feb 27 17:24:05 crc kubenswrapper[4814]: I0227 17:24:05.730975 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536884-lb8rx" Feb 27 17:24:05 crc kubenswrapper[4814]: I0227 17:24:05.795815 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfspp\" (UniqueName: \"kubernetes.io/projected/3596843e-9ae6-45eb-b1c9-e82215ced5e6-kube-api-access-pfspp\") pod \"3596843e-9ae6-45eb-b1c9-e82215ced5e6\" (UID: \"3596843e-9ae6-45eb-b1c9-e82215ced5e6\") " Feb 27 17:24:05 crc kubenswrapper[4814]: I0227 17:24:05.804930 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3596843e-9ae6-45eb-b1c9-e82215ced5e6-kube-api-access-pfspp" (OuterVolumeSpecName: "kube-api-access-pfspp") pod "3596843e-9ae6-45eb-b1c9-e82215ced5e6" (UID: "3596843e-9ae6-45eb-b1c9-e82215ced5e6"). InnerVolumeSpecName "kube-api-access-pfspp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:05 crc kubenswrapper[4814]: I0227 17:24:05.899764 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfspp\" (UniqueName: \"kubernetes.io/projected/3596843e-9ae6-45eb-b1c9-e82215ced5e6-kube-api-access-pfspp\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:06 crc kubenswrapper[4814]: I0227 17:24:06.364590 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536884-lb8rx" event={"ID":"3596843e-9ae6-45eb-b1c9-e82215ced5e6","Type":"ContainerDied","Data":"499b422985c0d4b559cbf794415df8f2a0d198354a7c4847716efe3386cc494a"} Feb 27 17:24:06 crc kubenswrapper[4814]: I0227 17:24:06.364651 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="499b422985c0d4b559cbf794415df8f2a0d198354a7c4847716efe3386cc494a" Feb 27 17:24:06 crc kubenswrapper[4814]: I0227 17:24:06.364717 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536884-lb8rx" Feb 27 17:24:06 crc kubenswrapper[4814]: I0227 17:24:06.436187 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536878-qb7qq"] Feb 27 17:24:06 crc kubenswrapper[4814]: I0227 17:24:06.446467 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536878-qb7qq"] Feb 27 17:24:06 crc kubenswrapper[4814]: I0227 17:24:06.506318 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdc6e071-8772-4c9e-811e-f2c747136f26" path="/var/lib/kubelet/pods/bdc6e071-8772-4c9e-811e-f2c747136f26/volumes" Feb 27 17:24:07 crc kubenswrapper[4814]: I0227 17:24:07.330536 4814 scope.go:117] "RemoveContainer" containerID="e6727f389a8f180b33335a320d5910e87674b4107164d84d2e352b0b35e6e40d" Feb 27 17:24:11 crc kubenswrapper[4814]: I0227 17:24:11.952766 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t6tx9/must-gather-mghpl"] Feb 27 17:24:11 crc kubenswrapper[4814]: E0227 17:24:11.954061 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3596843e-9ae6-45eb-b1c9-e82215ced5e6" containerName="oc" Feb 27 17:24:11 crc kubenswrapper[4814]: I0227 17:24:11.954075 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="3596843e-9ae6-45eb-b1c9-e82215ced5e6" containerName="oc" Feb 27 17:24:11 crc kubenswrapper[4814]: I0227 17:24:11.954333 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="3596843e-9ae6-45eb-b1c9-e82215ced5e6" containerName="oc" Feb 27 17:24:11 crc kubenswrapper[4814]: I0227 17:24:11.955414 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6tx9/must-gather-mghpl" Feb 27 17:24:11 crc kubenswrapper[4814]: I0227 17:24:11.959465 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-t6tx9"/"openshift-service-ca.crt" Feb 27 17:24:11 crc kubenswrapper[4814]: I0227 17:24:11.959709 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-t6tx9"/"default-dockercfg-rzxqx" Feb 27 17:24:11 crc kubenswrapper[4814]: I0227 17:24:11.959759 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-t6tx9"/"kube-root-ca.crt" Feb 27 17:24:11 crc kubenswrapper[4814]: I0227 17:24:11.971730 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-t6tx9/must-gather-mghpl"] Feb 27 17:24:12 crc kubenswrapper[4814]: I0227 17:24:12.156309 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgstx\" (UniqueName: \"kubernetes.io/projected/0e6b3bb7-c25b-491c-a582-a73eb2c15d85-kube-api-access-rgstx\") pod \"must-gather-mghpl\" (UID: \"0e6b3bb7-c25b-491c-a582-a73eb2c15d85\") " pod="openshift-must-gather-t6tx9/must-gather-mghpl" Feb 27 17:24:12 crc kubenswrapper[4814]: I0227 17:24:12.157479 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0e6b3bb7-c25b-491c-a582-a73eb2c15d85-must-gather-output\") pod \"must-gather-mghpl\" (UID: \"0e6b3bb7-c25b-491c-a582-a73eb2c15d85\") " pod="openshift-must-gather-t6tx9/must-gather-mghpl" Feb 27 17:24:12 crc kubenswrapper[4814]: I0227 17:24:12.259999 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgstx\" (UniqueName: \"kubernetes.io/projected/0e6b3bb7-c25b-491c-a582-a73eb2c15d85-kube-api-access-rgstx\") pod \"must-gather-mghpl\" (UID: \"0e6b3bb7-c25b-491c-a582-a73eb2c15d85\") " pod="openshift-must-gather-t6tx9/must-gather-mghpl" Feb 27 17:24:12 crc kubenswrapper[4814]: I0227 17:24:12.260299 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0e6b3bb7-c25b-491c-a582-a73eb2c15d85-must-gather-output\") pod \"must-gather-mghpl\" (UID: \"0e6b3bb7-c25b-491c-a582-a73eb2c15d85\") " pod="openshift-must-gather-t6tx9/must-gather-mghpl" Feb 27 17:24:12 crc kubenswrapper[4814]: I0227 17:24:12.261336 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0e6b3bb7-c25b-491c-a582-a73eb2c15d85-must-gather-output\") pod \"must-gather-mghpl\" (UID: \"0e6b3bb7-c25b-491c-a582-a73eb2c15d85\") " pod="openshift-must-gather-t6tx9/must-gather-mghpl" Feb 27 17:24:12 crc kubenswrapper[4814]: I0227 17:24:12.287227 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgstx\" (UniqueName: \"kubernetes.io/projected/0e6b3bb7-c25b-491c-a582-a73eb2c15d85-kube-api-access-rgstx\") pod \"must-gather-mghpl\" (UID: \"0e6b3bb7-c25b-491c-a582-a73eb2c15d85\") " pod="openshift-must-gather-t6tx9/must-gather-mghpl" Feb 27 17:24:12 crc kubenswrapper[4814]: I0227 17:24:12.580483 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6tx9/must-gather-mghpl" Feb 27 17:24:13 crc kubenswrapper[4814]: I0227 17:24:13.109165 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-t6tx9/must-gather-mghpl"] Feb 27 17:24:13 crc kubenswrapper[4814]: I0227 17:24:13.451602 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6tx9/must-gather-mghpl" event={"ID":"0e6b3bb7-c25b-491c-a582-a73eb2c15d85","Type":"ContainerStarted","Data":"a6f9f5628c111bb943250025d7ba63864bf5dd399554c9f4856c0972d6ee987d"} Feb 27 17:24:22 crc kubenswrapper[4814]: I0227 17:24:22.902905 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:24:22 crc kubenswrapper[4814]: I0227 17:24:22.904103 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:24:24 crc kubenswrapper[4814]: I0227 17:24:24.593988 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6tx9/must-gather-mghpl" event={"ID":"0e6b3bb7-c25b-491c-a582-a73eb2c15d85","Type":"ContainerStarted","Data":"f0d84c7eceeb93d66fcb4e5813c0721df388bd53a4f150d8e4a59287d656e14e"} Feb 27 17:24:25 crc kubenswrapper[4814]: I0227 17:24:25.611595 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6tx9/must-gather-mghpl" event={"ID":"0e6b3bb7-c25b-491c-a582-a73eb2c15d85","Type":"ContainerStarted","Data":"9d1bf312283009abb0f306d7c21ee78c73e613df948a46d0d87df7ed541b216f"} Feb 27 17:24:25 crc kubenswrapper[4814]: I0227 17:24:25.635527 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-t6tx9/must-gather-mghpl" podStartSLOduration=3.838379098 podStartE2EDuration="14.635493508s" podCreationTimestamp="2026-02-27 17:24:11 +0000 UTC" firstStartedPulling="2026-02-27 17:24:13.119347843 +0000 UTC m=+3665.571972713" lastFinishedPulling="2026-02-27 17:24:23.916462283 +0000 UTC m=+3676.369087123" observedRunningTime="2026-02-27 17:24:25.634168677 +0000 UTC m=+3678.086793507" watchObservedRunningTime="2026-02-27 17:24:25.635493508 +0000 UTC m=+3678.088118378" Feb 27 17:24:26 crc kubenswrapper[4814]: E0227 17:24:26.463056 4814 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.189:50394->38.102.83.189:38619: write tcp 38.102.83.189:50394->38.102.83.189:38619: write: broken pipe Feb 27 17:24:27 crc kubenswrapper[4814]: I0227 17:24:27.888534 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t6tx9/crc-debug-s5mnx"] Feb 27 17:24:27 crc kubenswrapper[4814]: I0227 17:24:27.891490 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6tx9/crc-debug-s5mnx" Feb 27 17:24:27 crc kubenswrapper[4814]: I0227 17:24:27.983940 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9-host\") pod \"crc-debug-s5mnx\" (UID: \"f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9\") " pod="openshift-must-gather-t6tx9/crc-debug-s5mnx" Feb 27 17:24:27 crc kubenswrapper[4814]: I0227 17:24:27.984015 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5r54\" (UniqueName: \"kubernetes.io/projected/f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9-kube-api-access-t5r54\") pod \"crc-debug-s5mnx\" (UID: \"f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9\") " pod="openshift-must-gather-t6tx9/crc-debug-s5mnx" Feb 27 17:24:28 crc kubenswrapper[4814]: I0227 17:24:28.086596 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9-host\") pod \"crc-debug-s5mnx\" (UID: \"f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9\") " pod="openshift-must-gather-t6tx9/crc-debug-s5mnx" Feb 27 17:24:28 crc kubenswrapper[4814]: I0227 17:24:28.086788 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9-host\") pod \"crc-debug-s5mnx\" (UID: \"f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9\") " pod="openshift-must-gather-t6tx9/crc-debug-s5mnx" Feb 27 17:24:28 crc kubenswrapper[4814]: I0227 17:24:28.087237 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5r54\" (UniqueName: \"kubernetes.io/projected/f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9-kube-api-access-t5r54\") pod \"crc-debug-s5mnx\" (UID: \"f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9\") " pod="openshift-must-gather-t6tx9/crc-debug-s5mnx" Feb 27 17:24:28 crc kubenswrapper[4814]: I0227 17:24:28.114390 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5r54\" (UniqueName: \"kubernetes.io/projected/f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9-kube-api-access-t5r54\") pod \"crc-debug-s5mnx\" (UID: \"f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9\") " pod="openshift-must-gather-t6tx9/crc-debug-s5mnx" Feb 27 17:24:28 crc kubenswrapper[4814]: I0227 17:24:28.220702 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6tx9/crc-debug-s5mnx" Feb 27 17:24:28 crc kubenswrapper[4814]: I0227 17:24:28.646547 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6tx9/crc-debug-s5mnx" event={"ID":"f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9","Type":"ContainerStarted","Data":"f8df69c257791c950097c60295b4558d2d74c11949698a335ca46c7c1d9891fb"} Feb 27 17:24:40 crc kubenswrapper[4814]: I0227 17:24:40.801140 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6tx9/crc-debug-s5mnx" event={"ID":"f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9","Type":"ContainerStarted","Data":"fdb18e92d8fbcf13ac997fc037379638ad73bc61cbd4c6d671c4b3a187d1f1f1"} Feb 27 17:24:40 crc kubenswrapper[4814]: I0227 17:24:40.826627 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-t6tx9/crc-debug-s5mnx" podStartSLOduration=2.5370568799999997 podStartE2EDuration="13.82660276s" podCreationTimestamp="2026-02-27 17:24:27 +0000 UTC" firstStartedPulling="2026-02-27 17:24:28.269090694 +0000 UTC m=+3680.721715524" lastFinishedPulling="2026-02-27 17:24:39.558636574 +0000 UTC m=+3692.011261404" observedRunningTime="2026-02-27 17:24:40.818300866 +0000 UTC m=+3693.270925696" watchObservedRunningTime="2026-02-27 17:24:40.82660276 +0000 UTC m=+3693.279227590" Feb 27 17:24:52 crc kubenswrapper[4814]: I0227 17:24:52.902615 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:24:52 crc kubenswrapper[4814]: I0227 17:24:52.903734 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:25:20 crc kubenswrapper[4814]: I0227 17:25:20.282113 4814 generic.go:334] "Generic (PLEG): container finished" podID="f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9" containerID="fdb18e92d8fbcf13ac997fc037379638ad73bc61cbd4c6d671c4b3a187d1f1f1" exitCode=0 Feb 27 17:25:20 crc kubenswrapper[4814]: I0227 17:25:20.283121 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6tx9/crc-debug-s5mnx" event={"ID":"f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9","Type":"ContainerDied","Data":"fdb18e92d8fbcf13ac997fc037379638ad73bc61cbd4c6d671c4b3a187d1f1f1"} Feb 27 17:25:21 crc kubenswrapper[4814]: I0227 17:25:21.396943 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6tx9/crc-debug-s5mnx" Feb 27 17:25:21 crc kubenswrapper[4814]: I0227 17:25:21.437296 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t6tx9/crc-debug-s5mnx"] Feb 27 17:25:21 crc kubenswrapper[4814]: I0227 17:25:21.448157 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t6tx9/crc-debug-s5mnx"] Feb 27 17:25:21 crc kubenswrapper[4814]: I0227 17:25:21.594964 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9-host\") pod \"f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9\" (UID: \"f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9\") " Feb 27 17:25:21 crc kubenswrapper[4814]: I0227 17:25:21.595477 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5r54\" (UniqueName: \"kubernetes.io/projected/f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9-kube-api-access-t5r54\") pod \"f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9\" (UID: \"f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9\") " Feb 27 17:25:21 crc kubenswrapper[4814]: I0227 17:25:21.595142 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9-host" (OuterVolumeSpecName: "host") pod "f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9" (UID: "f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:25:21 crc kubenswrapper[4814]: I0227 17:25:21.596638 4814 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9-host\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:21 crc kubenswrapper[4814]: I0227 17:25:21.603805 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9-kube-api-access-t5r54" (OuterVolumeSpecName: "kube-api-access-t5r54") pod "f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9" (UID: "f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9"). InnerVolumeSpecName "kube-api-access-t5r54". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:25:21 crc kubenswrapper[4814]: I0227 17:25:21.699607 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5r54\" (UniqueName: \"kubernetes.io/projected/f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9-kube-api-access-t5r54\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.306158 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8df69c257791c950097c60295b4558d2d74c11949698a335ca46c7c1d9891fb" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.306299 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6tx9/crc-debug-s5mnx" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.501853 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9" path="/var/lib/kubelet/pods/f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9/volumes" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.711651 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t6tx9/crc-debug-scplk"] Feb 27 17:25:22 crc kubenswrapper[4814]: E0227 17:25:22.712236 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9" containerName="container-00" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.712318 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9" containerName="container-00" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.712608 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="f76b9d4d-ca9f-45c2-a6bf-c6aeb8ba46d9" containerName="container-00" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.713498 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6tx9/crc-debug-scplk" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.723248 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx572\" (UniqueName: \"kubernetes.io/projected/a22d1910-571c-4518-a1f1-0d646eaa4d33-kube-api-access-cx572\") pod \"crc-debug-scplk\" (UID: \"a22d1910-571c-4518-a1f1-0d646eaa4d33\") " pod="openshift-must-gather-t6tx9/crc-debug-scplk" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.723425 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a22d1910-571c-4518-a1f1-0d646eaa4d33-host\") pod \"crc-debug-scplk\" (UID: \"a22d1910-571c-4518-a1f1-0d646eaa4d33\") " pod="openshift-must-gather-t6tx9/crc-debug-scplk" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.824748 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a22d1910-571c-4518-a1f1-0d646eaa4d33-host\") pod \"crc-debug-scplk\" (UID: \"a22d1910-571c-4518-a1f1-0d646eaa4d33\") " pod="openshift-must-gather-t6tx9/crc-debug-scplk" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.824851 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx572\" (UniqueName: \"kubernetes.io/projected/a22d1910-571c-4518-a1f1-0d646eaa4d33-kube-api-access-cx572\") pod \"crc-debug-scplk\" (UID: \"a22d1910-571c-4518-a1f1-0d646eaa4d33\") " pod="openshift-must-gather-t6tx9/crc-debug-scplk" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.824879 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a22d1910-571c-4518-a1f1-0d646eaa4d33-host\") pod \"crc-debug-scplk\" (UID: \"a22d1910-571c-4518-a1f1-0d646eaa4d33\") " pod="openshift-must-gather-t6tx9/crc-debug-scplk" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.850807 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx572\" (UniqueName: \"kubernetes.io/projected/a22d1910-571c-4518-a1f1-0d646eaa4d33-kube-api-access-cx572\") pod \"crc-debug-scplk\" (UID: \"a22d1910-571c-4518-a1f1-0d646eaa4d33\") " pod="openshift-must-gather-t6tx9/crc-debug-scplk" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.902569 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.902664 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.902737 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.904013 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"daa4ecfa76d33456eca88bbfe3cf3adb033c01a242837064c2ae775d74b120a7"} pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:25:22 crc kubenswrapper[4814]: I0227 17:25:22.904121 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" containerID="cri-o://daa4ecfa76d33456eca88bbfe3cf3adb033c01a242837064c2ae775d74b120a7" gracePeriod=600 Feb 27 17:25:23 crc kubenswrapper[4814]: I0227 17:25:23.031713 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6tx9/crc-debug-scplk" Feb 27 17:25:23 crc kubenswrapper[4814]: I0227 17:25:23.326236 4814 generic.go:334] "Generic (PLEG): container finished" podID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerID="daa4ecfa76d33456eca88bbfe3cf3adb033c01a242837064c2ae775d74b120a7" exitCode=0 Feb 27 17:25:23 crc kubenswrapper[4814]: I0227 17:25:23.326814 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerDied","Data":"daa4ecfa76d33456eca88bbfe3cf3adb033c01a242837064c2ae775d74b120a7"} Feb 27 17:25:23 crc kubenswrapper[4814]: I0227 17:25:23.326860 4814 scope.go:117] "RemoveContainer" containerID="422f090f169e341d48b95fe118355e130ce9821a26c4ba62c4f71b2744366f6f" Feb 27 17:25:23 crc kubenswrapper[4814]: I0227 17:25:23.330603 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6tx9/crc-debug-scplk" event={"ID":"a22d1910-571c-4518-a1f1-0d646eaa4d33","Type":"ContainerStarted","Data":"2177d6784e9cb095b7a8f406a9e001f719ac33547d4051e5d3d82d991138a4da"} Feb 27 17:25:24 crc kubenswrapper[4814]: I0227 17:25:24.347705 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3"} Feb 27 17:25:24 crc kubenswrapper[4814]: I0227 17:25:24.351322 4814 generic.go:334] "Generic (PLEG): container finished" podID="a22d1910-571c-4518-a1f1-0d646eaa4d33" containerID="fb641f670238671f623d98256379dea8bf6b61996aa684ce4b0285d944e322c0" exitCode=0 Feb 27 17:25:24 crc kubenswrapper[4814]: I0227 17:25:24.351388 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6tx9/crc-debug-scplk" event={"ID":"a22d1910-571c-4518-a1f1-0d646eaa4d33","Type":"ContainerDied","Data":"fb641f670238671f623d98256379dea8bf6b61996aa684ce4b0285d944e322c0"} Feb 27 17:25:24 crc kubenswrapper[4814]: I0227 17:25:24.851833 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t6tx9/crc-debug-scplk"] Feb 27 17:25:24 crc kubenswrapper[4814]: I0227 17:25:24.865688 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t6tx9/crc-debug-scplk"] Feb 27 17:25:25 crc kubenswrapper[4814]: I0227 17:25:25.483032 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6tx9/crc-debug-scplk" Feb 27 17:25:25 crc kubenswrapper[4814]: I0227 17:25:25.688617 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx572\" (UniqueName: \"kubernetes.io/projected/a22d1910-571c-4518-a1f1-0d646eaa4d33-kube-api-access-cx572\") pod \"a22d1910-571c-4518-a1f1-0d646eaa4d33\" (UID: \"a22d1910-571c-4518-a1f1-0d646eaa4d33\") " Feb 27 17:25:25 crc kubenswrapper[4814]: I0227 17:25:25.688871 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a22d1910-571c-4518-a1f1-0d646eaa4d33-host\") pod \"a22d1910-571c-4518-a1f1-0d646eaa4d33\" (UID: \"a22d1910-571c-4518-a1f1-0d646eaa4d33\") " Feb 27 17:25:25 crc kubenswrapper[4814]: I0227 17:25:25.690297 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a22d1910-571c-4518-a1f1-0d646eaa4d33-host" (OuterVolumeSpecName: "host") pod "a22d1910-571c-4518-a1f1-0d646eaa4d33" (UID: "a22d1910-571c-4518-a1f1-0d646eaa4d33"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:25:25 crc kubenswrapper[4814]: I0227 17:25:25.694796 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a22d1910-571c-4518-a1f1-0d646eaa4d33-kube-api-access-cx572" (OuterVolumeSpecName: "kube-api-access-cx572") pod "a22d1910-571c-4518-a1f1-0d646eaa4d33" (UID: "a22d1910-571c-4518-a1f1-0d646eaa4d33"). InnerVolumeSpecName "kube-api-access-cx572". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:25:25 crc kubenswrapper[4814]: I0227 17:25:25.791993 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx572\" (UniqueName: \"kubernetes.io/projected/a22d1910-571c-4518-a1f1-0d646eaa4d33-kube-api-access-cx572\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:25 crc kubenswrapper[4814]: I0227 17:25:25.792053 4814 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a22d1910-571c-4518-a1f1-0d646eaa4d33-host\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:26 crc kubenswrapper[4814]: I0227 17:25:26.036006 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t6tx9/crc-debug-bphph"] Feb 27 17:25:26 crc kubenswrapper[4814]: E0227 17:25:26.036477 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a22d1910-571c-4518-a1f1-0d646eaa4d33" containerName="container-00" Feb 27 17:25:26 crc kubenswrapper[4814]: I0227 17:25:26.036490 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="a22d1910-571c-4518-a1f1-0d646eaa4d33" containerName="container-00" Feb 27 17:25:26 crc kubenswrapper[4814]: I0227 17:25:26.036660 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="a22d1910-571c-4518-a1f1-0d646eaa4d33" containerName="container-00" Feb 27 17:25:26 crc kubenswrapper[4814]: I0227 17:25:26.037416 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6tx9/crc-debug-bphph" Feb 27 17:25:26 crc kubenswrapper[4814]: I0227 17:25:26.099711 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f9357f0a-2177-4c15-9f5d-daac9bac51f3-host\") pod \"crc-debug-bphph\" (UID: \"f9357f0a-2177-4c15-9f5d-daac9bac51f3\") " pod="openshift-must-gather-t6tx9/crc-debug-bphph" Feb 27 17:25:26 crc kubenswrapper[4814]: I0227 17:25:26.100409 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8zpp\" (UniqueName: \"kubernetes.io/projected/f9357f0a-2177-4c15-9f5d-daac9bac51f3-kube-api-access-z8zpp\") pod \"crc-debug-bphph\" (UID: \"f9357f0a-2177-4c15-9f5d-daac9bac51f3\") " pod="openshift-must-gather-t6tx9/crc-debug-bphph" Feb 27 17:25:26 crc kubenswrapper[4814]: I0227 17:25:26.203117 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f9357f0a-2177-4c15-9f5d-daac9bac51f3-host\") pod \"crc-debug-bphph\" (UID: \"f9357f0a-2177-4c15-9f5d-daac9bac51f3\") " pod="openshift-must-gather-t6tx9/crc-debug-bphph" Feb 27 17:25:26 crc kubenswrapper[4814]: I0227 17:25:26.203241 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f9357f0a-2177-4c15-9f5d-daac9bac51f3-host\") pod \"crc-debug-bphph\" (UID: \"f9357f0a-2177-4c15-9f5d-daac9bac51f3\") " pod="openshift-must-gather-t6tx9/crc-debug-bphph" Feb 27 17:25:26 crc kubenswrapper[4814]: I0227 17:25:26.203815 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8zpp\" (UniqueName: \"kubernetes.io/projected/f9357f0a-2177-4c15-9f5d-daac9bac51f3-kube-api-access-z8zpp\") pod \"crc-debug-bphph\" (UID: \"f9357f0a-2177-4c15-9f5d-daac9bac51f3\") " pod="openshift-must-gather-t6tx9/crc-debug-bphph" Feb 27 17:25:26 crc kubenswrapper[4814]: I0227 17:25:26.222809 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8zpp\" (UniqueName: \"kubernetes.io/projected/f9357f0a-2177-4c15-9f5d-daac9bac51f3-kube-api-access-z8zpp\") pod \"crc-debug-bphph\" (UID: \"f9357f0a-2177-4c15-9f5d-daac9bac51f3\") " pod="openshift-must-gather-t6tx9/crc-debug-bphph" Feb 27 17:25:26 crc kubenswrapper[4814]: I0227 17:25:26.353040 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6tx9/crc-debug-bphph" Feb 27 17:25:26 crc kubenswrapper[4814]: I0227 17:25:26.373545 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2177d6784e9cb095b7a8f406a9e001f719ac33547d4051e5d3d82d991138a4da" Feb 27 17:25:26 crc kubenswrapper[4814]: I0227 17:25:26.373772 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6tx9/crc-debug-scplk" Feb 27 17:25:26 crc kubenswrapper[4814]: W0227 17:25:26.398453 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9357f0a_2177_4c15_9f5d_daac9bac51f3.slice/crio-1f82b16803c88c1fb8c1a4d21976efe0506f1523b61aa4ebe387210f1d1b13d9 WatchSource:0}: Error finding container 1f82b16803c88c1fb8c1a4d21976efe0506f1523b61aa4ebe387210f1d1b13d9: Status 404 returned error can't find the container with id 1f82b16803c88c1fb8c1a4d21976efe0506f1523b61aa4ebe387210f1d1b13d9 Feb 27 17:25:26 crc kubenswrapper[4814]: I0227 17:25:26.503196 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a22d1910-571c-4518-a1f1-0d646eaa4d33" path="/var/lib/kubelet/pods/a22d1910-571c-4518-a1f1-0d646eaa4d33/volumes" Feb 27 17:25:27 crc kubenswrapper[4814]: I0227 17:25:27.393305 4814 generic.go:334] "Generic (PLEG): container finished" podID="f9357f0a-2177-4c15-9f5d-daac9bac51f3" containerID="e53d9d11f9b15cc68c6f9e63ddf2b9e5f06c09ca8707ece88b21e021cbf4f1c6" exitCode=0 Feb 27 17:25:27 crc kubenswrapper[4814]: I0227 17:25:27.393567 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6tx9/crc-debug-bphph" event={"ID":"f9357f0a-2177-4c15-9f5d-daac9bac51f3","Type":"ContainerDied","Data":"e53d9d11f9b15cc68c6f9e63ddf2b9e5f06c09ca8707ece88b21e021cbf4f1c6"} Feb 27 17:25:27 crc kubenswrapper[4814]: I0227 17:25:27.393807 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6tx9/crc-debug-bphph" event={"ID":"f9357f0a-2177-4c15-9f5d-daac9bac51f3","Type":"ContainerStarted","Data":"1f82b16803c88c1fb8c1a4d21976efe0506f1523b61aa4ebe387210f1d1b13d9"} Feb 27 17:25:27 crc kubenswrapper[4814]: I0227 17:25:27.453243 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t6tx9/crc-debug-bphph"] Feb 27 17:25:27 crc kubenswrapper[4814]: I0227 17:25:27.465760 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t6tx9/crc-debug-bphph"] Feb 27 17:25:28 crc kubenswrapper[4814]: I0227 17:25:28.543322 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6tx9/crc-debug-bphph" Feb 27 17:25:28 crc kubenswrapper[4814]: I0227 17:25:28.658590 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f9357f0a-2177-4c15-9f5d-daac9bac51f3-host\") pod \"f9357f0a-2177-4c15-9f5d-daac9bac51f3\" (UID: \"f9357f0a-2177-4c15-9f5d-daac9bac51f3\") " Feb 27 17:25:28 crc kubenswrapper[4814]: I0227 17:25:28.658739 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9357f0a-2177-4c15-9f5d-daac9bac51f3-host" (OuterVolumeSpecName: "host") pod "f9357f0a-2177-4c15-9f5d-daac9bac51f3" (UID: "f9357f0a-2177-4c15-9f5d-daac9bac51f3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:25:28 crc kubenswrapper[4814]: I0227 17:25:28.659158 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8zpp\" (UniqueName: \"kubernetes.io/projected/f9357f0a-2177-4c15-9f5d-daac9bac51f3-kube-api-access-z8zpp\") pod \"f9357f0a-2177-4c15-9f5d-daac9bac51f3\" (UID: \"f9357f0a-2177-4c15-9f5d-daac9bac51f3\") " Feb 27 17:25:28 crc kubenswrapper[4814]: I0227 17:25:28.660158 4814 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f9357f0a-2177-4c15-9f5d-daac9bac51f3-host\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:28 crc kubenswrapper[4814]: I0227 17:25:28.669451 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9357f0a-2177-4c15-9f5d-daac9bac51f3-kube-api-access-z8zpp" (OuterVolumeSpecName: "kube-api-access-z8zpp") pod "f9357f0a-2177-4c15-9f5d-daac9bac51f3" (UID: "f9357f0a-2177-4c15-9f5d-daac9bac51f3"). InnerVolumeSpecName "kube-api-access-z8zpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:25:28 crc kubenswrapper[4814]: I0227 17:25:28.762396 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8zpp\" (UniqueName: \"kubernetes.io/projected/f9357f0a-2177-4c15-9f5d-daac9bac51f3-kube-api-access-z8zpp\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:29 crc kubenswrapper[4814]: I0227 17:25:29.429985 4814 scope.go:117] "RemoveContainer" containerID="e53d9d11f9b15cc68c6f9e63ddf2b9e5f06c09ca8707ece88b21e021cbf4f1c6" Feb 27 17:25:29 crc kubenswrapper[4814]: I0227 17:25:29.430059 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6tx9/crc-debug-bphph" Feb 27 17:25:30 crc kubenswrapper[4814]: I0227 17:25:30.501223 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9357f0a-2177-4c15-9f5d-daac9bac51f3" path="/var/lib/kubelet/pods/f9357f0a-2177-4c15-9f5d-daac9bac51f3/volumes" Feb 27 17:25:44 crc kubenswrapper[4814]: I0227 17:25:44.981514 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-f8687c5bd-zdnxs_1121d7e8-a956-4a6b-929c-b1aecbaf16a4/barbican-api/0.log" Feb 27 17:25:45 crc kubenswrapper[4814]: I0227 17:25:45.193294 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-f8687c5bd-zdnxs_1121d7e8-a956-4a6b-929c-b1aecbaf16a4/barbican-api-log/0.log" Feb 27 17:25:45 crc kubenswrapper[4814]: I0227 17:25:45.232830 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6c5f495fc4-tkkrn_2bbc2d20-3264-487d-8571-71bc22fba348/barbican-keystone-listener/0.log" Feb 27 17:25:45 crc kubenswrapper[4814]: I0227 17:25:45.307663 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6c5f495fc4-tkkrn_2bbc2d20-3264-487d-8571-71bc22fba348/barbican-keystone-listener-log/0.log" Feb 27 17:25:45 crc kubenswrapper[4814]: I0227 17:25:45.460228 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-786585dcfc-sm2x8_b43b2496-ffec-4197-b0df-f03a3fc29ebf/barbican-worker/0.log" Feb 27 17:25:45 crc kubenswrapper[4814]: I0227 17:25:45.479769 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-786585dcfc-sm2x8_b43b2496-ffec-4197-b0df-f03a3fc29ebf/barbican-worker-log/0.log" Feb 27 17:25:45 crc kubenswrapper[4814]: I0227 17:25:45.684296 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs_556cb206-711d-4002-80cf-1ffe3b8f9643/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:25:45 crc kubenswrapper[4814]: I0227 17:25:45.733115 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_23a14569-9d74-4660-afec-2d91ae3a30a9/ceilometer-central-agent/0.log" Feb 27 17:25:45 crc kubenswrapper[4814]: I0227 17:25:45.772523 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_23a14569-9d74-4660-afec-2d91ae3a30a9/ceilometer-notification-agent/0.log" Feb 27 17:25:45 crc kubenswrapper[4814]: I0227 17:25:45.891032 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_23a14569-9d74-4660-afec-2d91ae3a30a9/proxy-httpd/0.log" Feb 27 17:25:45 crc kubenswrapper[4814]: I0227 17:25:45.947221 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_23a14569-9d74-4660-afec-2d91ae3a30a9/sg-core/0.log" Feb 27 17:25:45 crc kubenswrapper[4814]: I0227 17:25:45.996931 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_41da1428-a337-48a3-9609-35907e0d4955/cinder-api/0.log" Feb 27 17:25:46 crc kubenswrapper[4814]: I0227 17:25:46.119249 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_41da1428-a337-48a3-9609-35907e0d4955/cinder-api-log/0.log" Feb 27 17:25:46 crc kubenswrapper[4814]: I0227 17:25:46.247225 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_bd1ed9cd-1472-46c3-852f-a41a2a316e40/cinder-scheduler/0.log" Feb 27 17:25:46 crc kubenswrapper[4814]: I0227 17:25:46.248246 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_bd1ed9cd-1472-46c3-852f-a41a2a316e40/probe/0.log" Feb 27 17:25:46 crc kubenswrapper[4814]: I0227 17:25:46.583209 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw_1c0fc8c4-5559-4dd9-9220-b47c3b537749/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:25:46 crc kubenswrapper[4814]: I0227 17:25:46.661136 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-59k2x_3d27f572-8f1b-4507-a482-f3ce12031e76/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:25:46 crc kubenswrapper[4814]: I0227 17:25:46.785204 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-tww6q_18e28b72-72be-49fd-b568-c3a396e87db0/init/0.log" Feb 27 17:25:46 crc kubenswrapper[4814]: I0227 17:25:46.950959 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-tww6q_18e28b72-72be-49fd-b568-c3a396e87db0/init/0.log" Feb 27 17:25:47 crc kubenswrapper[4814]: I0227 17:25:47.041177 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-tww6q_18e28b72-72be-49fd-b568-c3a396e87db0/dnsmasq-dns/0.log" Feb 27 17:25:47 crc kubenswrapper[4814]: I0227 17:25:47.042991 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq_daf9ab93-de26-4929-a827-925d1650af5a/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:25:47 crc kubenswrapper[4814]: I0227 17:25:47.277152 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1e537a87-2e08-4521-bfa5-aeecdc14159d/glance-httpd/0.log" Feb 27 17:25:47 crc kubenswrapper[4814]: I0227 17:25:47.292135 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1e537a87-2e08-4521-bfa5-aeecdc14159d/glance-log/0.log" Feb 27 17:25:47 crc kubenswrapper[4814]: I0227 17:25:47.448535 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_74419238-7f7d-48d0-b138-f56913582843/glance-httpd/0.log" Feb 27 17:25:47 crc kubenswrapper[4814]: I0227 17:25:47.555376 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_74419238-7f7d-48d0-b138-f56913582843/glance-log/0.log" Feb 27 17:25:47 crc kubenswrapper[4814]: I0227 17:25:47.666381 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-65784f76f6-f2pcp_8af960c4-8a04-42d8-83bf-9d03c23ad333/horizon/0.log" Feb 27 17:25:47 crc kubenswrapper[4814]: I0227 17:25:47.811009 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-btx56_d9f3ab2c-0157-4a91-a329-c09648662cce/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:25:47 crc kubenswrapper[4814]: I0227 17:25:47.979381 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-65784f76f6-f2pcp_8af960c4-8a04-42d8-83bf-9d03c23ad333/horizon-log/0.log" Feb 27 17:25:48 crc kubenswrapper[4814]: I0227 17:25:48.014632 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-sqtqt_f4d8ff3a-2b21-45ab-8332-5bedf355c09c/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:25:48 crc kubenswrapper[4814]: I0227 17:25:48.297775 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29536861-pnnbc_24ebfded-b93e-4720-9747-75b8947d26c1/keystone-cron/0.log" Feb 27 17:25:48 crc kubenswrapper[4814]: I0227 17:25:48.340065 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7877c8f6d6-q56qs_017f253d-4956-4f30-93f1-66ead39f9980/keystone-api/0.log" Feb 27 17:25:48 crc kubenswrapper[4814]: I0227 17:25:48.462973 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_971d06e4-1037-4da5-b903-27ddda9e9f00/kube-state-metrics/0.log" Feb 27 17:25:48 crc kubenswrapper[4814]: I0227 17:25:48.753798 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n_c96b1d1f-6478-4eea-8413-a0657f19c6e7/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:25:49 crc kubenswrapper[4814]: I0227 17:25:49.029199 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88_cf70174d-7319-418e-8545-83db61a7eb7c/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:25:49 crc kubenswrapper[4814]: I0227 17:25:49.030945 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69895c5d4f-4bpj5_9e55471d-9b78-4e06-8195-c73f4ce0d1f6/neutron-api/0.log" Feb 27 17:25:49 crc kubenswrapper[4814]: I0227 17:25:49.045403 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69895c5d4f-4bpj5_9e55471d-9b78-4e06-8195-c73f4ce0d1f6/neutron-httpd/0.log" Feb 27 17:25:49 crc kubenswrapper[4814]: I0227 17:25:49.687968 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_1b31cce6-b000-448f-853e-231f707a2be5/nova-api-log/0.log" Feb 27 17:25:49 crc kubenswrapper[4814]: I0227 17:25:49.952239 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_f3152b8d-6ee7-4107-9205-b3535102b3cb/nova-cell0-conductor-conductor/0.log" Feb 27 17:25:50 crc kubenswrapper[4814]: I0227 17:25:50.004955 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_1b31cce6-b000-448f-853e-231f707a2be5/nova-api-api/0.log" Feb 27 17:25:50 crc kubenswrapper[4814]: I0227 17:25:50.042966 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_2bac854e-cfa6-4eaa-9272-5ecc27c88d5c/nova-cell1-conductor-conductor/0.log" Feb 27 17:25:50 crc kubenswrapper[4814]: I0227 17:25:50.316046 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_d4b56f88-0433-4ef4-b4d7-fec94f0f410a/nova-cell1-novncproxy-novncproxy/0.log" Feb 27 17:25:50 crc kubenswrapper[4814]: I0227 17:25:50.328123 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-48n4m_2f3a5797-c53e-4195-8e15-f2ba2016c410/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:25:50 crc kubenswrapper[4814]: I0227 17:25:50.682797 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b7902f26-5f7d-4c34-8166-d1857a631858/nova-metadata-log/0.log" Feb 27 17:25:50 crc kubenswrapper[4814]: I0227 17:25:50.784278 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_aace7eb3-3c5c-4290-8ada-b506bd913841/nova-scheduler-scheduler/0.log" Feb 27 17:25:50 crc kubenswrapper[4814]: I0227 17:25:50.922266 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ea6bae63-9ffa-4ba7-a95f-e686e8d12e87/mysql-bootstrap/0.log" Feb 27 17:25:51 crc kubenswrapper[4814]: I0227 17:25:51.169510 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ea6bae63-9ffa-4ba7-a95f-e686e8d12e87/mysql-bootstrap/0.log" Feb 27 17:25:51 crc kubenswrapper[4814]: I0227 17:25:51.176584 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ea6bae63-9ffa-4ba7-a95f-e686e8d12e87/galera/0.log" Feb 27 17:25:51 crc kubenswrapper[4814]: I0227 17:25:51.400904 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_502232b7-9e95-4215-9dfb-08054b866199/mysql-bootstrap/0.log" Feb 27 17:25:51 crc kubenswrapper[4814]: I0227 17:25:51.636432 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_502232b7-9e95-4215-9dfb-08054b866199/mysql-bootstrap/0.log" Feb 27 17:25:51 crc kubenswrapper[4814]: I0227 17:25:51.683306 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_502232b7-9e95-4215-9dfb-08054b866199/galera/0.log" Feb 27 17:25:51 crc kubenswrapper[4814]: I0227 17:25:51.833129 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b7902f26-5f7d-4c34-8166-d1857a631858/nova-metadata-metadata/0.log" Feb 27 17:25:51 crc kubenswrapper[4814]: I0227 17:25:51.880124 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_cc735347-66c8-4f38-89e9-31345e59ffee/openstackclient/0.log" Feb 27 17:25:52 crc kubenswrapper[4814]: I0227 17:25:52.059590 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-9tkjl_56c74238-2dde-4709-922a-9551da5fe8ae/ovn-controller/0.log" Feb 27 17:25:52 crc kubenswrapper[4814]: I0227 17:25:52.211272 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-hwlpr_b3a71748-678d-498b-8a4b-6bf26679c1cc/openstack-network-exporter/0.log" Feb 27 17:25:52 crc kubenswrapper[4814]: I0227 17:25:52.333910 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kxqxw_62963d98-fb43-4fb1-939a-6f8e6b69cb16/ovsdb-server-init/0.log" Feb 27 17:25:52 crc kubenswrapper[4814]: I0227 17:25:52.587958 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kxqxw_62963d98-fb43-4fb1-939a-6f8e6b69cb16/ovs-vswitchd/0.log" Feb 27 17:25:52 crc kubenswrapper[4814]: I0227 17:25:52.636734 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kxqxw_62963d98-fb43-4fb1-939a-6f8e6b69cb16/ovsdb-server/0.log" Feb 27 17:25:52 crc kubenswrapper[4814]: I0227 17:25:52.648433 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kxqxw_62963d98-fb43-4fb1-939a-6f8e6b69cb16/ovsdb-server-init/0.log" Feb 27 17:25:52 crc kubenswrapper[4814]: I0227 17:25:52.867975 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-58prd_245bd73e-7f47-4c71-9e26-c2b74c8b58e2/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:25:52 crc kubenswrapper[4814]: I0227 17:25:52.892818 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_381df711-c6a7-428a-94d5-8f2f84b51ef6/openstack-network-exporter/0.log" Feb 27 17:25:52 crc kubenswrapper[4814]: I0227 17:25:52.925591 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_381df711-c6a7-428a-94d5-8f2f84b51ef6/ovn-northd/0.log" Feb 27 17:25:53 crc kubenswrapper[4814]: I0227 17:25:53.291190 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_97ad2ba6-a67d-4926-93cd-1caf1c03e493/openstack-network-exporter/0.log" Feb 27 17:25:53 crc kubenswrapper[4814]: I0227 17:25:53.406688 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_97ad2ba6-a67d-4926-93cd-1caf1c03e493/ovsdbserver-nb/0.log" Feb 27 17:25:53 crc kubenswrapper[4814]: I0227 17:25:53.525396 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_bac59d25-01ed-44a7-9750-b654d1c7c631/ovsdbserver-sb/0.log" Feb 27 17:25:53 crc kubenswrapper[4814]: I0227 17:25:53.574015 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_bac59d25-01ed-44a7-9750-b654d1c7c631/openstack-network-exporter/0.log" Feb 27 17:25:53 crc kubenswrapper[4814]: I0227 17:25:53.808714 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8487496c64-cv9xr_13970ebf-9bd1-4292-a552-3f2d6bfa494c/placement-api/0.log" Feb 27 17:25:53 crc kubenswrapper[4814]: I0227 17:25:53.930105 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_51ca92ea-753f-4e44-94a8-68e73d165193/setup-container/0.log" Feb 27 17:25:53 crc kubenswrapper[4814]: I0227 17:25:53.965549 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8487496c64-cv9xr_13970ebf-9bd1-4292-a552-3f2d6bfa494c/placement-log/0.log" Feb 27 17:25:54 crc kubenswrapper[4814]: I0227 17:25:54.165303 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_51ca92ea-753f-4e44-94a8-68e73d165193/setup-container/0.log" Feb 27 17:25:54 crc kubenswrapper[4814]: I0227 17:25:54.187960 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_51ca92ea-753f-4e44-94a8-68e73d165193/rabbitmq/0.log" Feb 27 17:25:54 crc kubenswrapper[4814]: I0227 17:25:54.211969 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2fb1eae8-a753-43f3-8ed7-759e6e211ace/setup-container/0.log" Feb 27 17:25:54 crc kubenswrapper[4814]: I0227 17:25:54.516342 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2fb1eae8-a753-43f3-8ed7-759e6e211ace/setup-container/0.log" Feb 27 17:25:54 crc kubenswrapper[4814]: I0227 17:25:54.570852 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9_4cd3437d-9ecc-44b6-b9e7-7aaed24147a3/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:25:54 crc kubenswrapper[4814]: I0227 17:25:54.578787 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2fb1eae8-a753-43f3-8ed7-759e6e211ace/rabbitmq/0.log" Feb 27 17:25:54 crc kubenswrapper[4814]: I0227 17:25:54.774886 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-rqbpq_4d3976db-1cd1-4f9b-8975-038c57210a39/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:25:54 crc kubenswrapper[4814]: I0227 17:25:54.852227 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg_2bce282e-cd29-40eb-bb24-4ba8503d8a38/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:25:55 crc kubenswrapper[4814]: I0227 17:25:55.048138 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-s5rjk_223a940a-b149-4b9a-a328-411a730cf5fb/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:25:55 crc kubenswrapper[4814]: I0227 17:25:55.081991 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-kz77t_6fddd1b5-7d4c-4986-b192-f06f528c4b64/ssh-known-hosts-edpm-deployment/0.log" Feb 27 17:25:55 crc kubenswrapper[4814]: I0227 17:25:55.396821 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6d4ff56d55-bj24h_6cc6c3ab-803a-45de-a704-2e180b3bd2ce/proxy-server/0.log" Feb 27 17:25:55 crc kubenswrapper[4814]: I0227 17:25:55.429546 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6d4ff56d55-bj24h_6cc6c3ab-803a-45de-a704-2e180b3bd2ce/proxy-httpd/0.log" Feb 27 17:25:55 crc kubenswrapper[4814]: I0227 17:25:55.505384 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-k6ccx_3daf276e-d557-4192-831c-d2aa124fe9bc/swift-ring-rebalance/0.log" Feb 27 17:25:55 crc kubenswrapper[4814]: I0227 17:25:55.644931 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/account-auditor/0.log" Feb 27 17:25:55 crc kubenswrapper[4814]: I0227 17:25:55.705169 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/account-reaper/0.log" Feb 27 17:25:55 crc kubenswrapper[4814]: I0227 17:25:55.753790 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/account-replicator/0.log" Feb 27 17:25:55 crc kubenswrapper[4814]: I0227 17:25:55.901288 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/account-server/0.log" Feb 27 17:25:55 crc kubenswrapper[4814]: I0227 17:25:55.943117 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/container-auditor/0.log" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.009076 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/container-replicator/0.log" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.039776 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/container-server/0.log" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.133294 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/container-updater/0.log" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.271117 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t6tbf"] Feb 27 17:25:56 crc kubenswrapper[4814]: E0227 17:25:56.272232 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9357f0a-2177-4c15-9f5d-daac9bac51f3" containerName="container-00" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.272252 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9357f0a-2177-4c15-9f5d-daac9bac51f3" containerName="container-00" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.272553 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9357f0a-2177-4c15-9f5d-daac9bac51f3" containerName="container-00" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.274277 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.290979 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t6tbf"] Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.337887 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/object-auditor/0.log" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.374952 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/object-expirer/0.log" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.395714 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9pn4\" (UniqueName: \"kubernetes.io/projected/00a4027d-06c2-40ed-a137-d3580b7f5c0a-kube-api-access-r9pn4\") pod \"community-operators-t6tbf\" (UID: \"00a4027d-06c2-40ed-a137-d3580b7f5c0a\") " pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.395811 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a4027d-06c2-40ed-a137-d3580b7f5c0a-catalog-content\") pod \"community-operators-t6tbf\" (UID: \"00a4027d-06c2-40ed-a137-d3580b7f5c0a\") " pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.395919 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a4027d-06c2-40ed-a137-d3580b7f5c0a-utilities\") pod \"community-operators-t6tbf\" (UID: \"00a4027d-06c2-40ed-a137-d3580b7f5c0a\") " pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.413437 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/object-replicator/0.log" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.464591 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/object-server/0.log" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.497493 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9pn4\" (UniqueName: \"kubernetes.io/projected/00a4027d-06c2-40ed-a137-d3580b7f5c0a-kube-api-access-r9pn4\") pod \"community-operators-t6tbf\" (UID: \"00a4027d-06c2-40ed-a137-d3580b7f5c0a\") " pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.497545 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a4027d-06c2-40ed-a137-d3580b7f5c0a-catalog-content\") pod \"community-operators-t6tbf\" (UID: \"00a4027d-06c2-40ed-a137-d3580b7f5c0a\") " pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.497611 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a4027d-06c2-40ed-a137-d3580b7f5c0a-utilities\") pod \"community-operators-t6tbf\" (UID: \"00a4027d-06c2-40ed-a137-d3580b7f5c0a\") " pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.498224 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a4027d-06c2-40ed-a137-d3580b7f5c0a-utilities\") pod \"community-operators-t6tbf\" (UID: \"00a4027d-06c2-40ed-a137-d3580b7f5c0a\") " pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.498799 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a4027d-06c2-40ed-a137-d3580b7f5c0a-catalog-content\") pod \"community-operators-t6tbf\" (UID: \"00a4027d-06c2-40ed-a137-d3580b7f5c0a\") " pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.522372 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9pn4\" (UniqueName: \"kubernetes.io/projected/00a4027d-06c2-40ed-a137-d3580b7f5c0a-kube-api-access-r9pn4\") pod \"community-operators-t6tbf\" (UID: \"00a4027d-06c2-40ed-a137-d3580b7f5c0a\") " pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.613064 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:25:56 crc kubenswrapper[4814]: I0227 17:25:56.753904 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/object-updater/0.log" Feb 27 17:25:57 crc kubenswrapper[4814]: I0227 17:25:57.128543 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/swift-recon-cron/0.log" Feb 27 17:25:57 crc kubenswrapper[4814]: I0227 17:25:57.183456 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/rsync/0.log" Feb 27 17:25:57 crc kubenswrapper[4814]: I0227 17:25:57.402546 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-krg55_fee45495-5dbd-4679-b75d-bd5847b4a0fe/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:25:57 crc kubenswrapper[4814]: I0227 17:25:57.414793 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t6tbf"] Feb 27 17:25:57 crc kubenswrapper[4814]: I0227 17:25:57.434577 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_c11159aa-3255-4574-81a3-182c8a30d970/tempest-tests-tempest-tests-runner/0.log" Feb 27 17:25:57 crc kubenswrapper[4814]: I0227 17:25:57.717034 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_19e7467f-d219-4179-8f27-c3f0ecc2041f/test-operator-logs-container/0.log" Feb 27 17:25:57 crc kubenswrapper[4814]: I0227 17:25:57.720816 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-phrgr_4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:25:57 crc kubenswrapper[4814]: I0227 17:25:57.754749 4814 generic.go:334] "Generic (PLEG): container finished" podID="00a4027d-06c2-40ed-a137-d3580b7f5c0a" containerID="c7962397d25be67cfb42e700ecb33c3087b4ec35d0b9ba05989fe1d25e13b2ab" exitCode=0 Feb 27 17:25:57 crc kubenswrapper[4814]: I0227 17:25:57.754807 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6tbf" event={"ID":"00a4027d-06c2-40ed-a137-d3580b7f5c0a","Type":"ContainerDied","Data":"c7962397d25be67cfb42e700ecb33c3087b4ec35d0b9ba05989fe1d25e13b2ab"} Feb 27 17:25:57 crc kubenswrapper[4814]: I0227 17:25:57.754841 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6tbf" event={"ID":"00a4027d-06c2-40ed-a137-d3580b7f5c0a","Type":"ContainerStarted","Data":"1fa32dbdc6a11aeaea2862262fa15686b7fe7073093ed29eb7eb83fe9a22d24e"} Feb 27 17:25:59 crc kubenswrapper[4814]: I0227 17:25:59.777691 4814 generic.go:334] "Generic (PLEG): container finished" podID="00a4027d-06c2-40ed-a137-d3580b7f5c0a" containerID="4b23a2232d19274c620de4d6bd2b68a734f3eb1e6cb20d542033c69bb3fca76a" exitCode=0 Feb 27 17:25:59 crc kubenswrapper[4814]: I0227 17:25:59.777773 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6tbf" event={"ID":"00a4027d-06c2-40ed-a137-d3580b7f5c0a","Type":"ContainerDied","Data":"4b23a2232d19274c620de4d6bd2b68a734f3eb1e6cb20d542033c69bb3fca76a"} Feb 27 17:26:00 crc kubenswrapper[4814]: I0227 17:26:00.154730 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536886-pkn5g"] Feb 27 17:26:00 crc kubenswrapper[4814]: I0227 17:26:00.156376 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536886-pkn5g" Feb 27 17:26:00 crc kubenswrapper[4814]: I0227 17:26:00.159784 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:26:00 crc kubenswrapper[4814]: I0227 17:26:00.159947 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:26:00 crc kubenswrapper[4814]: I0227 17:26:00.159947 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:26:00 crc kubenswrapper[4814]: I0227 17:26:00.182437 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536886-pkn5g"] Feb 27 17:26:00 crc kubenswrapper[4814]: I0227 17:26:00.287190 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h592\" (UniqueName: \"kubernetes.io/projected/ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7-kube-api-access-6h592\") pod \"auto-csr-approver-29536886-pkn5g\" (UID: \"ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7\") " pod="openshift-infra/auto-csr-approver-29536886-pkn5g" Feb 27 17:26:00 crc kubenswrapper[4814]: I0227 17:26:00.389873 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h592\" (UniqueName: \"kubernetes.io/projected/ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7-kube-api-access-6h592\") pod \"auto-csr-approver-29536886-pkn5g\" (UID: \"ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7\") " pod="openshift-infra/auto-csr-approver-29536886-pkn5g" Feb 27 17:26:00 crc kubenswrapper[4814]: I0227 17:26:00.423047 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h592\" (UniqueName: \"kubernetes.io/projected/ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7-kube-api-access-6h592\") pod \"auto-csr-approver-29536886-pkn5g\" (UID: \"ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7\") " pod="openshift-infra/auto-csr-approver-29536886-pkn5g" Feb 27 17:26:00 crc kubenswrapper[4814]: I0227 17:26:00.476357 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536886-pkn5g" Feb 27 17:26:00 crc kubenswrapper[4814]: I0227 17:26:00.869565 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6tbf" event={"ID":"00a4027d-06c2-40ed-a137-d3580b7f5c0a","Type":"ContainerStarted","Data":"c4fa4925a4f98e9da1246aebe9007875e3130372d9de0fcd18da6e92ff95116d"} Feb 27 17:26:00 crc kubenswrapper[4814]: I0227 17:26:00.930558 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t6tbf" podStartSLOduration=2.5073007069999997 podStartE2EDuration="4.930525805s" podCreationTimestamp="2026-02-27 17:25:56 +0000 UTC" firstStartedPulling="2026-02-27 17:25:57.757291904 +0000 UTC m=+3770.209916734" lastFinishedPulling="2026-02-27 17:26:00.180517002 +0000 UTC m=+3772.633141832" observedRunningTime="2026-02-27 17:26:00.917132155 +0000 UTC m=+3773.369756985" watchObservedRunningTime="2026-02-27 17:26:00.930525805 +0000 UTC m=+3773.383150635" Feb 27 17:26:01 crc kubenswrapper[4814]: I0227 17:26:01.087103 4814 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:26:01 crc kubenswrapper[4814]: I0227 17:26:01.088390 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536886-pkn5g"] Feb 27 17:26:01 crc kubenswrapper[4814]: I0227 17:26:01.905924 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536886-pkn5g" event={"ID":"ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7","Type":"ContainerStarted","Data":"5faf41ffd1ac785256481a777bf3b68aeed6f015f4923b793b39eba9e5fe5bed"} Feb 27 17:26:02 crc kubenswrapper[4814]: I0227 17:26:02.915623 4814 generic.go:334] "Generic (PLEG): container finished" podID="ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7" containerID="f29f578709f1a3eb6fd8a6645dbd2766ce20ee9188e1615f824757758d72076d" exitCode=0 Feb 27 17:26:02 crc kubenswrapper[4814]: I0227 17:26:02.916037 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536886-pkn5g" event={"ID":"ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7","Type":"ContainerDied","Data":"f29f578709f1a3eb6fd8a6645dbd2766ce20ee9188e1615f824757758d72076d"} Feb 27 17:26:04 crc kubenswrapper[4814]: I0227 17:26:04.284198 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536886-pkn5g" Feb 27 17:26:04 crc kubenswrapper[4814]: I0227 17:26:04.380965 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h592\" (UniqueName: \"kubernetes.io/projected/ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7-kube-api-access-6h592\") pod \"ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7\" (UID: \"ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7\") " Feb 27 17:26:04 crc kubenswrapper[4814]: I0227 17:26:04.389305 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7-kube-api-access-6h592" (OuterVolumeSpecName: "kube-api-access-6h592") pod "ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7" (UID: "ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7"). InnerVolumeSpecName "kube-api-access-6h592". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:26:04 crc kubenswrapper[4814]: I0227 17:26:04.484306 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h592\" (UniqueName: \"kubernetes.io/projected/ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7-kube-api-access-6h592\") on node \"crc\" DevicePath \"\"" Feb 27 17:26:04 crc kubenswrapper[4814]: I0227 17:26:04.956972 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536886-pkn5g" event={"ID":"ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7","Type":"ContainerDied","Data":"5faf41ffd1ac785256481a777bf3b68aeed6f015f4923b793b39eba9e5fe5bed"} Feb 27 17:26:04 crc kubenswrapper[4814]: I0227 17:26:04.957347 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5faf41ffd1ac785256481a777bf3b68aeed6f015f4923b793b39eba9e5fe5bed" Feb 27 17:26:04 crc kubenswrapper[4814]: I0227 17:26:04.957462 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536886-pkn5g" Feb 27 17:26:05 crc kubenswrapper[4814]: I0227 17:26:05.403650 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536880-pbzpg"] Feb 27 17:26:05 crc kubenswrapper[4814]: I0227 17:26:05.425913 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536880-pbzpg"] Feb 27 17:26:06 crc kubenswrapper[4814]: I0227 17:26:06.498291 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42a61257-51ec-49cd-9e0f-3f872cc8a93b" path="/var/lib/kubelet/pods/42a61257-51ec-49cd-9e0f-3f872cc8a93b/volumes" Feb 27 17:26:06 crc kubenswrapper[4814]: I0227 17:26:06.613433 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:26:06 crc kubenswrapper[4814]: I0227 17:26:06.613644 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:26:06 crc kubenswrapper[4814]: I0227 17:26:06.664411 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:26:07 crc kubenswrapper[4814]: I0227 17:26:07.036588 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:26:07 crc kubenswrapper[4814]: I0227 17:26:07.042808 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_414562e2-0c3b-4db9-bc71-78928f882398/memcached/0.log" Feb 27 17:26:07 crc kubenswrapper[4814]: I0227 17:26:07.095202 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t6tbf"] Feb 27 17:26:07 crc kubenswrapper[4814]: I0227 17:26:07.497868 4814 scope.go:117] "RemoveContainer" containerID="de34d4f24044ae6e3fb99b5c983f19f17262db1f14fdb116f3cb03b0c052bbd1" Feb 27 17:26:08 crc kubenswrapper[4814]: I0227 17:26:08.997449 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t6tbf" podUID="00a4027d-06c2-40ed-a137-d3580b7f5c0a" containerName="registry-server" containerID="cri-o://c4fa4925a4f98e9da1246aebe9007875e3130372d9de0fcd18da6e92ff95116d" gracePeriod=2 Feb 27 17:26:09 crc kubenswrapper[4814]: I0227 17:26:09.556737 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:26:09 crc kubenswrapper[4814]: I0227 17:26:09.589517 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a4027d-06c2-40ed-a137-d3580b7f5c0a-utilities\") pod \"00a4027d-06c2-40ed-a137-d3580b7f5c0a\" (UID: \"00a4027d-06c2-40ed-a137-d3580b7f5c0a\") " Feb 27 17:26:09 crc kubenswrapper[4814]: I0227 17:26:09.589604 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9pn4\" (UniqueName: \"kubernetes.io/projected/00a4027d-06c2-40ed-a137-d3580b7f5c0a-kube-api-access-r9pn4\") pod \"00a4027d-06c2-40ed-a137-d3580b7f5c0a\" (UID: \"00a4027d-06c2-40ed-a137-d3580b7f5c0a\") " Feb 27 17:26:09 crc kubenswrapper[4814]: I0227 17:26:09.589741 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a4027d-06c2-40ed-a137-d3580b7f5c0a-catalog-content\") pod \"00a4027d-06c2-40ed-a137-d3580b7f5c0a\" (UID: \"00a4027d-06c2-40ed-a137-d3580b7f5c0a\") " Feb 27 17:26:09 crc kubenswrapper[4814]: I0227 17:26:09.590596 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00a4027d-06c2-40ed-a137-d3580b7f5c0a-utilities" (OuterVolumeSpecName: "utilities") pod "00a4027d-06c2-40ed-a137-d3580b7f5c0a" (UID: "00a4027d-06c2-40ed-a137-d3580b7f5c0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:26:09 crc kubenswrapper[4814]: I0227 17:26:09.597399 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00a4027d-06c2-40ed-a137-d3580b7f5c0a-kube-api-access-r9pn4" (OuterVolumeSpecName: "kube-api-access-r9pn4") pod "00a4027d-06c2-40ed-a137-d3580b7f5c0a" (UID: "00a4027d-06c2-40ed-a137-d3580b7f5c0a"). InnerVolumeSpecName "kube-api-access-r9pn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:26:09 crc kubenswrapper[4814]: I0227 17:26:09.693591 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a4027d-06c2-40ed-a137-d3580b7f5c0a-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:26:09 crc kubenswrapper[4814]: I0227 17:26:09.693645 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9pn4\" (UniqueName: \"kubernetes.io/projected/00a4027d-06c2-40ed-a137-d3580b7f5c0a-kube-api-access-r9pn4\") on node \"crc\" DevicePath \"\"" Feb 27 17:26:09 crc kubenswrapper[4814]: I0227 17:26:09.858799 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00a4027d-06c2-40ed-a137-d3580b7f5c0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00a4027d-06c2-40ed-a137-d3580b7f5c0a" (UID: "00a4027d-06c2-40ed-a137-d3580b7f5c0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:26:09 crc kubenswrapper[4814]: I0227 17:26:09.897760 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a4027d-06c2-40ed-a137-d3580b7f5c0a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:26:10 crc kubenswrapper[4814]: I0227 17:26:10.021340 4814 generic.go:334] "Generic (PLEG): container finished" podID="00a4027d-06c2-40ed-a137-d3580b7f5c0a" containerID="c4fa4925a4f98e9da1246aebe9007875e3130372d9de0fcd18da6e92ff95116d" exitCode=0 Feb 27 17:26:10 crc kubenswrapper[4814]: I0227 17:26:10.021380 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6tbf" event={"ID":"00a4027d-06c2-40ed-a137-d3580b7f5c0a","Type":"ContainerDied","Data":"c4fa4925a4f98e9da1246aebe9007875e3130372d9de0fcd18da6e92ff95116d"} Feb 27 17:26:10 crc kubenswrapper[4814]: I0227 17:26:10.021417 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t6tbf" Feb 27 17:26:10 crc kubenswrapper[4814]: I0227 17:26:10.021445 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6tbf" event={"ID":"00a4027d-06c2-40ed-a137-d3580b7f5c0a","Type":"ContainerDied","Data":"1fa32dbdc6a11aeaea2862262fa15686b7fe7073093ed29eb7eb83fe9a22d24e"} Feb 27 17:26:10 crc kubenswrapper[4814]: I0227 17:26:10.021472 4814 scope.go:117] "RemoveContainer" containerID="c4fa4925a4f98e9da1246aebe9007875e3130372d9de0fcd18da6e92ff95116d" Feb 27 17:26:10 crc kubenswrapper[4814]: I0227 17:26:10.050483 4814 scope.go:117] "RemoveContainer" containerID="4b23a2232d19274c620de4d6bd2b68a734f3eb1e6cb20d542033c69bb3fca76a" Feb 27 17:26:10 crc kubenswrapper[4814]: I0227 17:26:10.058834 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t6tbf"] Feb 27 17:26:10 crc kubenswrapper[4814]: I0227 17:26:10.069709 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t6tbf"] Feb 27 17:26:10 crc kubenswrapper[4814]: I0227 17:26:10.080149 4814 scope.go:117] "RemoveContainer" containerID="c7962397d25be67cfb42e700ecb33c3087b4ec35d0b9ba05989fe1d25e13b2ab" Feb 27 17:26:10 crc kubenswrapper[4814]: I0227 17:26:10.117111 4814 scope.go:117] "RemoveContainer" containerID="c4fa4925a4f98e9da1246aebe9007875e3130372d9de0fcd18da6e92ff95116d" Feb 27 17:26:10 crc kubenswrapper[4814]: E0227 17:26:10.117841 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4fa4925a4f98e9da1246aebe9007875e3130372d9de0fcd18da6e92ff95116d\": container with ID starting with c4fa4925a4f98e9da1246aebe9007875e3130372d9de0fcd18da6e92ff95116d not found: ID does not exist" containerID="c4fa4925a4f98e9da1246aebe9007875e3130372d9de0fcd18da6e92ff95116d" Feb 27 17:26:10 crc kubenswrapper[4814]: I0227 17:26:10.117899 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4fa4925a4f98e9da1246aebe9007875e3130372d9de0fcd18da6e92ff95116d"} err="failed to get container status \"c4fa4925a4f98e9da1246aebe9007875e3130372d9de0fcd18da6e92ff95116d\": rpc error: code = NotFound desc = could not find container \"c4fa4925a4f98e9da1246aebe9007875e3130372d9de0fcd18da6e92ff95116d\": container with ID starting with c4fa4925a4f98e9da1246aebe9007875e3130372d9de0fcd18da6e92ff95116d not found: ID does not exist" Feb 27 17:26:10 crc kubenswrapper[4814]: I0227 17:26:10.117929 4814 scope.go:117] "RemoveContainer" containerID="4b23a2232d19274c620de4d6bd2b68a734f3eb1e6cb20d542033c69bb3fca76a" Feb 27 17:26:10 crc kubenswrapper[4814]: E0227 17:26:10.118372 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b23a2232d19274c620de4d6bd2b68a734f3eb1e6cb20d542033c69bb3fca76a\": container with ID starting with 4b23a2232d19274c620de4d6bd2b68a734f3eb1e6cb20d542033c69bb3fca76a not found: ID does not exist" containerID="4b23a2232d19274c620de4d6bd2b68a734f3eb1e6cb20d542033c69bb3fca76a" Feb 27 17:26:10 crc kubenswrapper[4814]: I0227 17:26:10.118397 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b23a2232d19274c620de4d6bd2b68a734f3eb1e6cb20d542033c69bb3fca76a"} err="failed to get container status \"4b23a2232d19274c620de4d6bd2b68a734f3eb1e6cb20d542033c69bb3fca76a\": rpc error: code = NotFound desc = could not find container \"4b23a2232d19274c620de4d6bd2b68a734f3eb1e6cb20d542033c69bb3fca76a\": container with ID starting with 4b23a2232d19274c620de4d6bd2b68a734f3eb1e6cb20d542033c69bb3fca76a not found: ID does not exist" Feb 27 17:26:10 crc kubenswrapper[4814]: I0227 17:26:10.118412 4814 scope.go:117] "RemoveContainer" containerID="c7962397d25be67cfb42e700ecb33c3087b4ec35d0b9ba05989fe1d25e13b2ab" Feb 27 17:26:10 crc kubenswrapper[4814]: E0227 17:26:10.118832 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7962397d25be67cfb42e700ecb33c3087b4ec35d0b9ba05989fe1d25e13b2ab\": container with ID starting with c7962397d25be67cfb42e700ecb33c3087b4ec35d0b9ba05989fe1d25e13b2ab not found: ID does not exist" containerID="c7962397d25be67cfb42e700ecb33c3087b4ec35d0b9ba05989fe1d25e13b2ab" Feb 27 17:26:10 crc kubenswrapper[4814]: I0227 17:26:10.118862 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7962397d25be67cfb42e700ecb33c3087b4ec35d0b9ba05989fe1d25e13b2ab"} err="failed to get container status \"c7962397d25be67cfb42e700ecb33c3087b4ec35d0b9ba05989fe1d25e13b2ab\": rpc error: code = NotFound desc = could not find container \"c7962397d25be67cfb42e700ecb33c3087b4ec35d0b9ba05989fe1d25e13b2ab\": container with ID starting with c7962397d25be67cfb42e700ecb33c3087b4ec35d0b9ba05989fe1d25e13b2ab not found: ID does not exist" Feb 27 17:26:10 crc kubenswrapper[4814]: I0227 17:26:10.499730 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00a4027d-06c2-40ed-a137-d3580b7f5c0a" path="/var/lib/kubelet/pods/00a4027d-06c2-40ed-a137-d3580b7f5c0a/volumes" Feb 27 17:26:26 crc kubenswrapper[4814]: I0227 17:26:26.570110 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g_0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8/util/0.log" Feb 27 17:26:26 crc kubenswrapper[4814]: I0227 17:26:26.773114 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g_0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8/util/0.log" Feb 27 17:26:26 crc kubenswrapper[4814]: I0227 17:26:26.790954 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g_0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8/pull/0.log" Feb 27 17:26:26 crc kubenswrapper[4814]: I0227 17:26:26.884133 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g_0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8/pull/0.log" Feb 27 17:26:26 crc kubenswrapper[4814]: I0227 17:26:26.995097 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g_0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8/pull/0.log" Feb 27 17:26:26 crc kubenswrapper[4814]: I0227 17:26:26.997210 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g_0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8/util/0.log" Feb 27 17:26:27 crc kubenswrapper[4814]: I0227 17:26:27.120050 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g_0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8/extract/0.log" Feb 27 17:26:27 crc kubenswrapper[4814]: I0227 17:26:27.441924 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-5d87c9d997-pjncz_6ee0f0c6-52ee-4d20-999f-a0a579dcdaef/manager/0.log" Feb 27 17:26:27 crc kubenswrapper[4814]: I0227 17:26:27.825883 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-64db6967f8-9kqgh_49b03138-9b4c-486c-af91-37c16e8c2536/manager/0.log" Feb 27 17:26:28 crc kubenswrapper[4814]: I0227 17:26:28.002299 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-cf99c678f-ccrlj_a3038d2f-e4de-4d0f-8ac0-e5c666077395/manager/0.log" Feb 27 17:26:28 crc kubenswrapper[4814]: I0227 17:26:28.264076 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-78bc7f9bd9-p74f5_ea6ecef0-2796-4c32-9a9a-ee178574d2f9/manager/0.log" Feb 27 17:26:28 crc kubenswrapper[4814]: I0227 17:26:28.577440 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-55d77d7b5c-2d2n7_de6e87db-4283-4aab-a97e-23792fd3b480/manager/0.log" Feb 27 17:26:28 crc kubenswrapper[4814]: I0227 17:26:28.695827 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-545456dc4-fp474_2192fb73-bd5d-440f-8c01-2871658164d9/manager/0.log" Feb 27 17:26:28 crc kubenswrapper[4814]: I0227 17:26:28.882976 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-f7fcc58b9-qc48n_2d989d20-0a79-448b-8b73-ad9378119232/manager/0.log" Feb 27 17:26:29 crc kubenswrapper[4814]: I0227 17:26:29.018589 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55ffd4876b-nxrtn_9b633770-3463-40ee-b5fc-48ec5fb4846c/manager/0.log" Feb 27 17:26:29 crc kubenswrapper[4814]: I0227 17:26:29.112085 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-67d996989d-c5mcj_dfcfffcb-36da-4853-8ed7-ba76caba9d3c/manager/0.log" Feb 27 17:26:29 crc kubenswrapper[4814]: I0227 17:26:29.373202 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-556b8b874-f4dv9_71ebdc25-566e-4b42-8f48-6f3c52e1125c/manager/0.log" Feb 27 17:26:29 crc kubenswrapper[4814]: I0227 17:26:29.566377 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54688575f-q2gkq_50d18356-e250-46b6-bcd3-053f770f8f58/manager/0.log" Feb 27 17:26:29 crc kubenswrapper[4814]: I0227 17:26:29.709783 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-74b6b5dc96-m5zqv_5ee5efe1-368e-476e-8516-d9b81d9a38a6/manager/0.log" Feb 27 17:26:29 crc kubenswrapper[4814]: I0227 17:26:29.892729 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5d86c7ddb7-4m2wv_75ee2f25-f59a-4a60-8245-1b638957d234/manager/0.log" Feb 27 17:26:30 crc kubenswrapper[4814]: I0227 17:26:30.030119 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr_edf51635-71a3-4c03-8aef-50212a3b8247/manager/0.log" Feb 27 17:26:30 crc kubenswrapper[4814]: I0227 17:26:30.341347 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-7d5f5556b6-b9hqd_43af9a66-30c9-4410-862d-e07c38625ea7/operator/0.log" Feb 27 17:26:30 crc kubenswrapper[4814]: I0227 17:26:30.465512 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-rnx2c_f571f89f-5528-4a9c-9ea4-f1449e67eb56/registry-server/0.log" Feb 27 17:26:30 crc kubenswrapper[4814]: I0227 17:26:30.631599 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-75684d597f-5958h_c6faf190-d834-45e6-8a18-067067c0bef5/manager/0.log" Feb 27 17:26:30 crc kubenswrapper[4814]: I0227 17:26:30.797696 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-648564c9fc-hbdxk_103a95f2-68f7-41b0-9dea-46cde1679842/manager/0.log" Feb 27 17:26:30 crc kubenswrapper[4814]: I0227 17:26:30.979673 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-lwz52_60ad038c-aad7-4309-89cf-17bf1bda00ac/operator/0.log" Feb 27 17:26:31 crc kubenswrapper[4814]: I0227 17:26:31.084158 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9b9ff9f4d-pq6mn_59cb5f17-969c-4c4f-81bf-a5b0e65acace/manager/0.log" Feb 27 17:26:31 crc kubenswrapper[4814]: I0227 17:26:31.332028 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5fdb694969-4s2gh_3f85fa1d-a7c6-49b8-8f85-d02670fa2849/manager/0.log" Feb 27 17:26:31 crc kubenswrapper[4814]: I0227 17:26:31.442532 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-55b5ff4dbb-tck96_30b75b23-8950-40e1-a515-eaf136867a25/manager/0.log" Feb 27 17:26:31 crc kubenswrapper[4814]: I0227 17:26:31.569784 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-bccc79885-hh455_4b02bae2-c5ac-423e-a2f1-b95de1264f47/manager/0.log" Feb 27 17:26:31 crc kubenswrapper[4814]: I0227 17:26:31.973326 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-655dd5d898-d7f5g_96d3c178-c0f0-4bd5-b3bd-0e572404d30a/manager/0.log" Feb 27 17:26:33 crc kubenswrapper[4814]: I0227 17:26:33.076779 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6db6876945-xnhhg_6ede514a-7a77-4cfe-a746-3d7bd2f5d7fd/manager/0.log" Feb 27 17:26:54 crc kubenswrapper[4814]: I0227 17:26:54.046670 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-ws5nw_3dafc1d6-a249-4d30-ad93-f065110076c8/control-plane-machine-set-operator/0.log" Feb 27 17:26:54 crc kubenswrapper[4814]: I0227 17:26:54.206128 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ndc8k_bc099c2d-2785-4509-bd8b-29469341f56d/kube-rbac-proxy/0.log" Feb 27 17:26:54 crc kubenswrapper[4814]: I0227 17:26:54.227321 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ndc8k_bc099c2d-2785-4509-bd8b-29469341f56d/machine-api-operator/0.log" Feb 27 17:27:09 crc kubenswrapper[4814]: I0227 17:27:09.226220 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-wnnxz_9043722e-d6cc-495d-bcf6-ad5b95545ca6/cert-manager-controller/0.log" Feb 27 17:27:09 crc kubenswrapper[4814]: I0227 17:27:09.474767 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-77gdd_c6ddbede-41e8-4a88-8348-a81b242a9b85/cert-manager-cainjector/0.log" Feb 27 17:27:09 crc kubenswrapper[4814]: I0227 17:27:09.492397 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-lpq28_b97362bd-6925-4d89-aba9-a8ad9a9cb1bb/cert-manager-webhook/0.log" Feb 27 17:27:25 crc kubenswrapper[4814]: I0227 17:27:25.181776 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5dcbbd79cf-5hqsn_ec633892-636c-460c-8c7a-1b594ba270b5/nmstate-console-plugin/0.log" Feb 27 17:27:25 crc kubenswrapper[4814]: I0227 17:27:25.313671 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-6j6pw_694e9d8b-01d0-444c-bf6f-cc4c3cd4f542/nmstate-handler/0.log" Feb 27 17:27:25 crc kubenswrapper[4814]: I0227 17:27:25.387019 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-69594cc75-9xknj_6df39380-a69c-4ef3-94ba-6a5ef31e7f50/kube-rbac-proxy/0.log" Feb 27 17:27:25 crc kubenswrapper[4814]: I0227 17:27:25.426550 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-69594cc75-9xknj_6df39380-a69c-4ef3-94ba-6a5ef31e7f50/nmstate-metrics/0.log" Feb 27 17:27:25 crc kubenswrapper[4814]: I0227 17:27:25.532987 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-75c5dccd6c-r6vt9_154b3bf1-f2f5-4e4b-9110-b2097784f5d8/nmstate-operator/0.log" Feb 27 17:27:25 crc kubenswrapper[4814]: I0227 17:27:25.616426 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-786f45cff4-bbbjp_65270046-4e69-4b5c-b07f-3f401949f32b/nmstate-webhook/0.log" Feb 27 17:27:52 crc kubenswrapper[4814]: I0227 17:27:52.902430 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:27:52 crc kubenswrapper[4814]: I0227 17:27:52.903361 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:27:58 crc kubenswrapper[4814]: I0227 17:27:58.186925 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-5lqjv_a197efea-0275-49bc-acd5-b043737cd1a8/kube-rbac-proxy/0.log" Feb 27 17:27:58 crc kubenswrapper[4814]: I0227 17:27:58.345785 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-5lqjv_a197efea-0275-49bc-acd5-b043737cd1a8/controller/0.log" Feb 27 17:27:58 crc kubenswrapper[4814]: I0227 17:27:58.401598 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-frr-files/0.log" Feb 27 17:27:58 crc kubenswrapper[4814]: I0227 17:27:58.612561 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-reloader/0.log" Feb 27 17:27:58 crc kubenswrapper[4814]: I0227 17:27:58.629570 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-frr-files/0.log" Feb 27 17:27:58 crc kubenswrapper[4814]: I0227 17:27:58.658522 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-reloader/0.log" Feb 27 17:27:58 crc kubenswrapper[4814]: I0227 17:27:58.667839 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-metrics/0.log" Feb 27 17:27:58 crc kubenswrapper[4814]: I0227 17:27:58.877063 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-reloader/0.log" Feb 27 17:27:58 crc kubenswrapper[4814]: I0227 17:27:58.879525 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-metrics/0.log" Feb 27 17:27:58 crc kubenswrapper[4814]: I0227 17:27:58.888850 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-frr-files/0.log" Feb 27 17:27:58 crc kubenswrapper[4814]: I0227 17:27:58.910731 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-metrics/0.log" Feb 27 17:27:59 crc kubenswrapper[4814]: I0227 17:27:59.102871 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-frr-files/0.log" Feb 27 17:27:59 crc kubenswrapper[4814]: I0227 17:27:59.103707 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-reloader/0.log" Feb 27 17:27:59 crc kubenswrapper[4814]: I0227 17:27:59.147723 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-metrics/0.log" Feb 27 17:27:59 crc kubenswrapper[4814]: I0227 17:27:59.170710 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/controller/0.log" Feb 27 17:27:59 crc kubenswrapper[4814]: I0227 17:27:59.335539 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/frr-metrics/0.log" Feb 27 17:27:59 crc kubenswrapper[4814]: I0227 17:27:59.375506 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/kube-rbac-proxy/0.log" Feb 27 17:27:59 crc kubenswrapper[4814]: I0227 17:27:59.414152 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/kube-rbac-proxy-frr/0.log" Feb 27 17:27:59 crc kubenswrapper[4814]: I0227 17:27:59.588022 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/reloader/0.log" Feb 27 17:27:59 crc kubenswrapper[4814]: I0227 17:27:59.690366 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7f989f654f-lt84m_bbc12f0a-9041-484e-9216-9a618b63257b/frr-k8s-webhook-server/0.log" Feb 27 17:27:59 crc kubenswrapper[4814]: I0227 17:27:59.899465 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-58945cb6fc-nth4f_e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9/manager/0.log" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.068653 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5d5d4f645d-nwdxv_20a110bc-385c-4cb3-a3e6-ce3568069d69/webhook-server/0.log" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.159816 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536888-mcrll"] Feb 27 17:28:00 crc kubenswrapper[4814]: E0227 17:28:00.160434 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a4027d-06c2-40ed-a137-d3580b7f5c0a" containerName="registry-server" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.160454 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a4027d-06c2-40ed-a137-d3580b7f5c0a" containerName="registry-server" Feb 27 17:28:00 crc kubenswrapper[4814]: E0227 17:28:00.160478 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7" containerName="oc" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.160485 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7" containerName="oc" Feb 27 17:28:00 crc kubenswrapper[4814]: E0227 17:28:00.160497 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a4027d-06c2-40ed-a137-d3580b7f5c0a" containerName="extract-utilities" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.160504 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a4027d-06c2-40ed-a137-d3580b7f5c0a" containerName="extract-utilities" Feb 27 17:28:00 crc kubenswrapper[4814]: E0227 17:28:00.160538 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a4027d-06c2-40ed-a137-d3580b7f5c0a" containerName="extract-content" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.160544 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a4027d-06c2-40ed-a137-d3580b7f5c0a" containerName="extract-content" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.160737 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="00a4027d-06c2-40ed-a137-d3580b7f5c0a" containerName="registry-server" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.160764 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7" containerName="oc" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.161585 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536888-mcrll" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.166591 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.166663 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.166883 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.172607 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536888-mcrll"] Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.204614 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-v59vj_44587c86-594d-410c-939c-f3a1426ae15a/kube-rbac-proxy/0.log" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.300845 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqr5p\" (UniqueName: \"kubernetes.io/projected/fbea8566-2a73-43f6-9b57-8372114149fb-kube-api-access-tqr5p\") pod \"auto-csr-approver-29536888-mcrll\" (UID: \"fbea8566-2a73-43f6-9b57-8372114149fb\") " pod="openshift-infra/auto-csr-approver-29536888-mcrll" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.404881 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqr5p\" (UniqueName: \"kubernetes.io/projected/fbea8566-2a73-43f6-9b57-8372114149fb-kube-api-access-tqr5p\") pod \"auto-csr-approver-29536888-mcrll\" (UID: \"fbea8566-2a73-43f6-9b57-8372114149fb\") " pod="openshift-infra/auto-csr-approver-29536888-mcrll" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.429120 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqr5p\" (UniqueName: \"kubernetes.io/projected/fbea8566-2a73-43f6-9b57-8372114149fb-kube-api-access-tqr5p\") pod \"auto-csr-approver-29536888-mcrll\" (UID: \"fbea8566-2a73-43f6-9b57-8372114149fb\") " pod="openshift-infra/auto-csr-approver-29536888-mcrll" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.485511 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536888-mcrll" Feb 27 17:28:00 crc kubenswrapper[4814]: I0227 17:28:00.905692 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536888-mcrll"] Feb 27 17:28:01 crc kubenswrapper[4814]: I0227 17:28:01.018927 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-v59vj_44587c86-594d-410c-939c-f3a1426ae15a/speaker/0.log" Feb 27 17:28:01 crc kubenswrapper[4814]: I0227 17:28:01.108846 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/frr/0.log" Feb 27 17:28:01 crc kubenswrapper[4814]: I0227 17:28:01.243967 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536888-mcrll" event={"ID":"fbea8566-2a73-43f6-9b57-8372114149fb","Type":"ContainerStarted","Data":"09a285ec70513cd49950446d6bbf5e0768c53da45e6a9eed0fb04988bac177c8"} Feb 27 17:28:02 crc kubenswrapper[4814]: I0227 17:28:02.264480 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536888-mcrll" event={"ID":"fbea8566-2a73-43f6-9b57-8372114149fb","Type":"ContainerStarted","Data":"e5e285ac2f7eaff66116ffcb22e228c0831edba91d524019e361e3bdfedaa28b"} Feb 27 17:28:02 crc kubenswrapper[4814]: I0227 17:28:02.292144 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536888-mcrll" podStartSLOduration=1.395044214 podStartE2EDuration="2.292115999s" podCreationTimestamp="2026-02-27 17:28:00 +0000 UTC" firstStartedPulling="2026-02-27 17:28:00.941537157 +0000 UTC m=+3893.394161987" lastFinishedPulling="2026-02-27 17:28:01.838608942 +0000 UTC m=+3894.291233772" observedRunningTime="2026-02-27 17:28:02.280677234 +0000 UTC m=+3894.733302064" watchObservedRunningTime="2026-02-27 17:28:02.292115999 +0000 UTC m=+3894.744740829" Feb 27 17:28:03 crc kubenswrapper[4814]: I0227 17:28:03.279100 4814 generic.go:334] "Generic (PLEG): container finished" podID="fbea8566-2a73-43f6-9b57-8372114149fb" containerID="e5e285ac2f7eaff66116ffcb22e228c0831edba91d524019e361e3bdfedaa28b" exitCode=0 Feb 27 17:28:03 crc kubenswrapper[4814]: I0227 17:28:03.279214 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536888-mcrll" event={"ID":"fbea8566-2a73-43f6-9b57-8372114149fb","Type":"ContainerDied","Data":"e5e285ac2f7eaff66116ffcb22e228c0831edba91d524019e361e3bdfedaa28b"} Feb 27 17:28:04 crc kubenswrapper[4814]: I0227 17:28:04.688031 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536888-mcrll" Feb 27 17:28:04 crc kubenswrapper[4814]: I0227 17:28:04.823638 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqr5p\" (UniqueName: \"kubernetes.io/projected/fbea8566-2a73-43f6-9b57-8372114149fb-kube-api-access-tqr5p\") pod \"fbea8566-2a73-43f6-9b57-8372114149fb\" (UID: \"fbea8566-2a73-43f6-9b57-8372114149fb\") " Feb 27 17:28:04 crc kubenswrapper[4814]: I0227 17:28:04.832604 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbea8566-2a73-43f6-9b57-8372114149fb-kube-api-access-tqr5p" (OuterVolumeSpecName: "kube-api-access-tqr5p") pod "fbea8566-2a73-43f6-9b57-8372114149fb" (UID: "fbea8566-2a73-43f6-9b57-8372114149fb"). InnerVolumeSpecName "kube-api-access-tqr5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:28:04 crc kubenswrapper[4814]: I0227 17:28:04.926435 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqr5p\" (UniqueName: \"kubernetes.io/projected/fbea8566-2a73-43f6-9b57-8372114149fb-kube-api-access-tqr5p\") on node \"crc\" DevicePath \"\"" Feb 27 17:28:05 crc kubenswrapper[4814]: I0227 17:28:05.310566 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536888-mcrll" event={"ID":"fbea8566-2a73-43f6-9b57-8372114149fb","Type":"ContainerDied","Data":"09a285ec70513cd49950446d6bbf5e0768c53da45e6a9eed0fb04988bac177c8"} Feb 27 17:28:05 crc kubenswrapper[4814]: I0227 17:28:05.311083 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09a285ec70513cd49950446d6bbf5e0768c53da45e6a9eed0fb04988bac177c8" Feb 27 17:28:05 crc kubenswrapper[4814]: I0227 17:28:05.310636 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536888-mcrll" Feb 27 17:28:05 crc kubenswrapper[4814]: I0227 17:28:05.377961 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536882-pj72p"] Feb 27 17:28:05 crc kubenswrapper[4814]: I0227 17:28:05.389489 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536882-pj72p"] Feb 27 17:28:06 crc kubenswrapper[4814]: I0227 17:28:06.504705 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16bef484-7466-490c-80ec-55001c91d531" path="/var/lib/kubelet/pods/16bef484-7466-490c-80ec-55001c91d531/volumes" Feb 27 17:28:07 crc kubenswrapper[4814]: I0227 17:28:07.651965 4814 scope.go:117] "RemoveContainer" containerID="d2708a39c6e6743ab5bad4a16e91735a1af99b89f2e978a63babb49df312533d" Feb 27 17:28:15 crc kubenswrapper[4814]: I0227 17:28:15.009140 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lr9lx"] Feb 27 17:28:15 crc kubenswrapper[4814]: E0227 17:28:15.010496 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbea8566-2a73-43f6-9b57-8372114149fb" containerName="oc" Feb 27 17:28:15 crc kubenswrapper[4814]: I0227 17:28:15.010514 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbea8566-2a73-43f6-9b57-8372114149fb" containerName="oc" Feb 27 17:28:15 crc kubenswrapper[4814]: I0227 17:28:15.010766 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbea8566-2a73-43f6-9b57-8372114149fb" containerName="oc" Feb 27 17:28:15 crc kubenswrapper[4814]: I0227 17:28:15.012315 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:15 crc kubenswrapper[4814]: I0227 17:28:15.029511 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lr9lx"] Feb 27 17:28:15 crc kubenswrapper[4814]: I0227 17:28:15.190954 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-catalog-content\") pod \"redhat-marketplace-lr9lx\" (UID: \"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10\") " pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:15 crc kubenswrapper[4814]: I0227 17:28:15.191049 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt52d\" (UniqueName: \"kubernetes.io/projected/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-kube-api-access-rt52d\") pod \"redhat-marketplace-lr9lx\" (UID: \"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10\") " pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:15 crc kubenswrapper[4814]: I0227 17:28:15.191692 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-utilities\") pod \"redhat-marketplace-lr9lx\" (UID: \"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10\") " pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:15 crc kubenswrapper[4814]: I0227 17:28:15.293722 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt52d\" (UniqueName: \"kubernetes.io/projected/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-kube-api-access-rt52d\") pod \"redhat-marketplace-lr9lx\" (UID: \"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10\") " pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:15 crc kubenswrapper[4814]: I0227 17:28:15.293934 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-utilities\") pod \"redhat-marketplace-lr9lx\" (UID: \"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10\") " pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:15 crc kubenswrapper[4814]: I0227 17:28:15.294005 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-catalog-content\") pod \"redhat-marketplace-lr9lx\" (UID: \"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10\") " pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:15 crc kubenswrapper[4814]: I0227 17:28:15.294626 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-catalog-content\") pod \"redhat-marketplace-lr9lx\" (UID: \"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10\") " pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:15 crc kubenswrapper[4814]: I0227 17:28:15.294870 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-utilities\") pod \"redhat-marketplace-lr9lx\" (UID: \"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10\") " pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:15 crc kubenswrapper[4814]: I0227 17:28:15.318662 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt52d\" (UniqueName: \"kubernetes.io/projected/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-kube-api-access-rt52d\") pod \"redhat-marketplace-lr9lx\" (UID: \"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10\") " pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:15 crc kubenswrapper[4814]: I0227 17:28:15.333500 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:15 crc kubenswrapper[4814]: I0227 17:28:15.819874 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lr9lx"] Feb 27 17:28:16 crc kubenswrapper[4814]: I0227 17:28:16.431787 4814 generic.go:334] "Generic (PLEG): container finished" podID="ae394f60-eb2c-4b59-ba67-bb7a3e34ef10" containerID="95864b6707e4e5b1d402ade41538fbc4d4bf54f88e5aa4792d270cc5dcf91992" exitCode=0 Feb 27 17:28:16 crc kubenswrapper[4814]: I0227 17:28:16.432286 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lr9lx" event={"ID":"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10","Type":"ContainerDied","Data":"95864b6707e4e5b1d402ade41538fbc4d4bf54f88e5aa4792d270cc5dcf91992"} Feb 27 17:28:16 crc kubenswrapper[4814]: I0227 17:28:16.432328 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lr9lx" event={"ID":"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10","Type":"ContainerStarted","Data":"ca42e974dba710d59781fe73b99e7e6698817bb8de270ead398401e40d80d194"} Feb 27 17:28:16 crc kubenswrapper[4814]: I0227 17:28:16.987573 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr_4fb39e04-954d-4a82-9d19-ece0f11c5118/util/0.log" Feb 27 17:28:17 crc kubenswrapper[4814]: I0227 17:28:17.249504 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr_4fb39e04-954d-4a82-9d19-ece0f11c5118/util/0.log" Feb 27 17:28:17 crc kubenswrapper[4814]: I0227 17:28:17.250157 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr_4fb39e04-954d-4a82-9d19-ece0f11c5118/pull/0.log" Feb 27 17:28:17 crc kubenswrapper[4814]: I0227 17:28:17.264364 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr_4fb39e04-954d-4a82-9d19-ece0f11c5118/pull/0.log" Feb 27 17:28:17 crc kubenswrapper[4814]: I0227 17:28:17.406717 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr_4fb39e04-954d-4a82-9d19-ece0f11c5118/util/0.log" Feb 27 17:28:17 crc kubenswrapper[4814]: I0227 17:28:17.456925 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr_4fb39e04-954d-4a82-9d19-ece0f11c5118/pull/0.log" Feb 27 17:28:17 crc kubenswrapper[4814]: I0227 17:28:17.466261 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr_4fb39e04-954d-4a82-9d19-ece0f11c5118/extract/0.log" Feb 27 17:28:17 crc kubenswrapper[4814]: I0227 17:28:17.655044 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kbbd8_705d8544-aed0-40dd-aac2-185cfc8ecb06/extract-utilities/0.log" Feb 27 17:28:17 crc kubenswrapper[4814]: I0227 17:28:17.816616 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kbbd8_705d8544-aed0-40dd-aac2-185cfc8ecb06/extract-content/0.log" Feb 27 17:28:17 crc kubenswrapper[4814]: I0227 17:28:17.839006 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kbbd8_705d8544-aed0-40dd-aac2-185cfc8ecb06/extract-content/0.log" Feb 27 17:28:17 crc kubenswrapper[4814]: I0227 17:28:17.840429 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kbbd8_705d8544-aed0-40dd-aac2-185cfc8ecb06/extract-utilities/0.log" Feb 27 17:28:18 crc kubenswrapper[4814]: I0227 17:28:18.062208 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kbbd8_705d8544-aed0-40dd-aac2-185cfc8ecb06/extract-content/0.log" Feb 27 17:28:18 crc kubenswrapper[4814]: I0227 17:28:18.076284 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kbbd8_705d8544-aed0-40dd-aac2-185cfc8ecb06/extract-utilities/0.log" Feb 27 17:28:18 crc kubenswrapper[4814]: I0227 17:28:18.241186 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kbbd8_705d8544-aed0-40dd-aac2-185cfc8ecb06/registry-server/0.log" Feb 27 17:28:18 crc kubenswrapper[4814]: I0227 17:28:18.310571 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4nmp_5452c803-ae28-4bd0-a404-5f3d707b14d5/extract-utilities/0.log" Feb 27 17:28:18 crc kubenswrapper[4814]: I0227 17:28:18.452080 4814 generic.go:334] "Generic (PLEG): container finished" podID="ae394f60-eb2c-4b59-ba67-bb7a3e34ef10" containerID="61821c5666cd74e8a1ef8c012e45d23be1cb6eaed2afc33cf8b6d161824d6fc7" exitCode=0 Feb 27 17:28:18 crc kubenswrapper[4814]: I0227 17:28:18.452148 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lr9lx" event={"ID":"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10","Type":"ContainerDied","Data":"61821c5666cd74e8a1ef8c012e45d23be1cb6eaed2afc33cf8b6d161824d6fc7"} Feb 27 17:28:18 crc kubenswrapper[4814]: I0227 17:28:18.487672 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4nmp_5452c803-ae28-4bd0-a404-5f3d707b14d5/extract-content/0.log" Feb 27 17:28:18 crc kubenswrapper[4814]: I0227 17:28:18.535604 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4nmp_5452c803-ae28-4bd0-a404-5f3d707b14d5/extract-content/0.log" Feb 27 17:28:18 crc kubenswrapper[4814]: I0227 17:28:18.558009 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4nmp_5452c803-ae28-4bd0-a404-5f3d707b14d5/extract-utilities/0.log" Feb 27 17:28:18 crc kubenswrapper[4814]: I0227 17:28:18.731566 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4nmp_5452c803-ae28-4bd0-a404-5f3d707b14d5/extract-utilities/0.log" Feb 27 17:28:18 crc kubenswrapper[4814]: I0227 17:28:18.767554 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4nmp_5452c803-ae28-4bd0-a404-5f3d707b14d5/extract-content/0.log" Feb 27 17:28:19 crc kubenswrapper[4814]: I0227 17:28:19.025501 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj_b790a0b4-30f0-40fa-9c06-6c91abe332e0/util/0.log" Feb 27 17:28:19 crc kubenswrapper[4814]: I0227 17:28:19.471153 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj_b790a0b4-30f0-40fa-9c06-6c91abe332e0/util/0.log" Feb 27 17:28:19 crc kubenswrapper[4814]: I0227 17:28:19.486910 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lr9lx" event={"ID":"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10","Type":"ContainerStarted","Data":"9cdbbb4d34288d7e54be431e0d3772b144177ad70c33076be31027f900812046"} Feb 27 17:28:19 crc kubenswrapper[4814]: I0227 17:28:19.508891 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lr9lx" podStartSLOduration=3.025302323 podStartE2EDuration="5.508868593s" podCreationTimestamp="2026-02-27 17:28:14 +0000 UTC" firstStartedPulling="2026-02-27 17:28:16.437185628 +0000 UTC m=+3908.889810458" lastFinishedPulling="2026-02-27 17:28:18.920751898 +0000 UTC m=+3911.373376728" observedRunningTime="2026-02-27 17:28:19.504684332 +0000 UTC m=+3911.957309152" watchObservedRunningTime="2026-02-27 17:28:19.508868593 +0000 UTC m=+3911.961493423" Feb 27 17:28:19 crc kubenswrapper[4814]: I0227 17:28:19.527480 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj_b790a0b4-30f0-40fa-9c06-6c91abe332e0/pull/0.log" Feb 27 17:28:19 crc kubenswrapper[4814]: I0227 17:28:19.529643 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj_b790a0b4-30f0-40fa-9c06-6c91abe332e0/pull/0.log" Feb 27 17:28:19 crc kubenswrapper[4814]: I0227 17:28:19.561977 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4nmp_5452c803-ae28-4bd0-a404-5f3d707b14d5/registry-server/0.log" Feb 27 17:28:19 crc kubenswrapper[4814]: I0227 17:28:19.786225 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj_b790a0b4-30f0-40fa-9c06-6c91abe332e0/pull/0.log" Feb 27 17:28:19 crc kubenswrapper[4814]: I0227 17:28:19.804510 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj_b790a0b4-30f0-40fa-9c06-6c91abe332e0/extract/0.log" Feb 27 17:28:19 crc kubenswrapper[4814]: I0227 17:28:19.813007 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj_b790a0b4-30f0-40fa-9c06-6c91abe332e0/util/0.log" Feb 27 17:28:19 crc kubenswrapper[4814]: I0227 17:28:19.982832 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-lvrr5_c5d324fb-1565-4d40-9304-9e2f6adc9604/marketplace-operator/0.log" Feb 27 17:28:20 crc kubenswrapper[4814]: I0227 17:28:20.073971 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fdwvd_01865c94-699f-4fe9-8427-bfe3a2788531/extract-utilities/0.log" Feb 27 17:28:20 crc kubenswrapper[4814]: I0227 17:28:20.270102 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fdwvd_01865c94-699f-4fe9-8427-bfe3a2788531/extract-utilities/0.log" Feb 27 17:28:20 crc kubenswrapper[4814]: I0227 17:28:20.298441 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fdwvd_01865c94-699f-4fe9-8427-bfe3a2788531/extract-content/0.log" Feb 27 17:28:20 crc kubenswrapper[4814]: I0227 17:28:20.335537 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fdwvd_01865c94-699f-4fe9-8427-bfe3a2788531/extract-content/0.log" Feb 27 17:28:20 crc kubenswrapper[4814]: I0227 17:28:20.559515 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fdwvd_01865c94-699f-4fe9-8427-bfe3a2788531/extract-utilities/0.log" Feb 27 17:28:20 crc kubenswrapper[4814]: I0227 17:28:20.574707 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fdwvd_01865c94-699f-4fe9-8427-bfe3a2788531/extract-content/0.log" Feb 27 17:28:20 crc kubenswrapper[4814]: I0227 17:28:20.635709 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fdwvd_01865c94-699f-4fe9-8427-bfe3a2788531/registry-server/0.log" Feb 27 17:28:20 crc kubenswrapper[4814]: I0227 17:28:20.770449 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lr9lx_ae394f60-eb2c-4b59-ba67-bb7a3e34ef10/extract-utilities/0.log" Feb 27 17:28:20 crc kubenswrapper[4814]: I0227 17:28:20.999687 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lr9lx_ae394f60-eb2c-4b59-ba67-bb7a3e34ef10/extract-utilities/0.log" Feb 27 17:28:21 crc kubenswrapper[4814]: I0227 17:28:21.005231 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lr9lx_ae394f60-eb2c-4b59-ba67-bb7a3e34ef10/extract-content/0.log" Feb 27 17:28:21 crc kubenswrapper[4814]: I0227 17:28:21.045437 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lr9lx_ae394f60-eb2c-4b59-ba67-bb7a3e34ef10/extract-content/0.log" Feb 27 17:28:21 crc kubenswrapper[4814]: I0227 17:28:21.190060 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lr9lx_ae394f60-eb2c-4b59-ba67-bb7a3e34ef10/extract-content/0.log" Feb 27 17:28:21 crc kubenswrapper[4814]: I0227 17:28:21.221830 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lr9lx_ae394f60-eb2c-4b59-ba67-bb7a3e34ef10/registry-server/0.log" Feb 27 17:28:21 crc kubenswrapper[4814]: I0227 17:28:21.226932 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lr9lx_ae394f60-eb2c-4b59-ba67-bb7a3e34ef10/extract-utilities/0.log" Feb 27 17:28:21 crc kubenswrapper[4814]: I0227 17:28:21.274938 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5rrh_73643f07-8d1a-4321-902c-12a52a2fe4da/extract-utilities/0.log" Feb 27 17:28:21 crc kubenswrapper[4814]: I0227 17:28:21.451320 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5rrh_73643f07-8d1a-4321-902c-12a52a2fe4da/extract-utilities/0.log" Feb 27 17:28:21 crc kubenswrapper[4814]: I0227 17:28:21.459692 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5rrh_73643f07-8d1a-4321-902c-12a52a2fe4da/extract-content/0.log" Feb 27 17:28:21 crc kubenswrapper[4814]: I0227 17:28:21.482855 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5rrh_73643f07-8d1a-4321-902c-12a52a2fe4da/extract-content/0.log" Feb 27 17:28:21 crc kubenswrapper[4814]: I0227 17:28:21.649165 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5rrh_73643f07-8d1a-4321-902c-12a52a2fe4da/extract-utilities/0.log" Feb 27 17:28:21 crc kubenswrapper[4814]: I0227 17:28:21.684757 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5rrh_73643f07-8d1a-4321-902c-12a52a2fe4da/extract-content/0.log" Feb 27 17:28:22 crc kubenswrapper[4814]: I0227 17:28:22.163923 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5rrh_73643f07-8d1a-4321-902c-12a52a2fe4da/registry-server/0.log" Feb 27 17:28:22 crc kubenswrapper[4814]: I0227 17:28:22.902485 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:28:22 crc kubenswrapper[4814]: I0227 17:28:22.902592 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:28:25 crc kubenswrapper[4814]: I0227 17:28:25.333727 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:25 crc kubenswrapper[4814]: I0227 17:28:25.334209 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:25 crc kubenswrapper[4814]: I0227 17:28:25.402046 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:25 crc kubenswrapper[4814]: I0227 17:28:25.603850 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:25 crc kubenswrapper[4814]: I0227 17:28:25.665510 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lr9lx"] Feb 27 17:28:27 crc kubenswrapper[4814]: I0227 17:28:27.574481 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lr9lx" podUID="ae394f60-eb2c-4b59-ba67-bb7a3e34ef10" containerName="registry-server" containerID="cri-o://9cdbbb4d34288d7e54be431e0d3772b144177ad70c33076be31027f900812046" gracePeriod=2 Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.031299 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.133781 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt52d\" (UniqueName: \"kubernetes.io/projected/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-kube-api-access-rt52d\") pod \"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10\" (UID: \"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10\") " Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.133981 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-utilities\") pod \"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10\" (UID: \"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10\") " Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.134059 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-catalog-content\") pod \"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10\" (UID: \"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10\") " Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.136567 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-utilities" (OuterVolumeSpecName: "utilities") pod "ae394f60-eb2c-4b59-ba67-bb7a3e34ef10" (UID: "ae394f60-eb2c-4b59-ba67-bb7a3e34ef10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.167312 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-kube-api-access-rt52d" (OuterVolumeSpecName: "kube-api-access-rt52d") pod "ae394f60-eb2c-4b59-ba67-bb7a3e34ef10" (UID: "ae394f60-eb2c-4b59-ba67-bb7a3e34ef10"). InnerVolumeSpecName "kube-api-access-rt52d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.197029 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae394f60-eb2c-4b59-ba67-bb7a3e34ef10" (UID: "ae394f60-eb2c-4b59-ba67-bb7a3e34ef10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.256086 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt52d\" (UniqueName: \"kubernetes.io/projected/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-kube-api-access-rt52d\") on node \"crc\" DevicePath \"\"" Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.256137 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.256148 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.586108 4814 generic.go:334] "Generic (PLEG): container finished" podID="ae394f60-eb2c-4b59-ba67-bb7a3e34ef10" containerID="9cdbbb4d34288d7e54be431e0d3772b144177ad70c33076be31027f900812046" exitCode=0 Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.586166 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lr9lx" event={"ID":"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10","Type":"ContainerDied","Data":"9cdbbb4d34288d7e54be431e0d3772b144177ad70c33076be31027f900812046"} Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.586204 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lr9lx" event={"ID":"ae394f60-eb2c-4b59-ba67-bb7a3e34ef10","Type":"ContainerDied","Data":"ca42e974dba710d59781fe73b99e7e6698817bb8de270ead398401e40d80d194"} Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.586228 4814 scope.go:117] "RemoveContainer" containerID="9cdbbb4d34288d7e54be431e0d3772b144177ad70c33076be31027f900812046" Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.586263 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lr9lx" Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.611624 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lr9lx"] Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.617326 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lr9lx"] Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.625652 4814 scope.go:117] "RemoveContainer" containerID="61821c5666cd74e8a1ef8c012e45d23be1cb6eaed2afc33cf8b6d161824d6fc7" Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.663747 4814 scope.go:117] "RemoveContainer" containerID="95864b6707e4e5b1d402ade41538fbc4d4bf54f88e5aa4792d270cc5dcf91992" Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.692781 4814 scope.go:117] "RemoveContainer" containerID="9cdbbb4d34288d7e54be431e0d3772b144177ad70c33076be31027f900812046" Feb 27 17:28:28 crc kubenswrapper[4814]: E0227 17:28:28.693128 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cdbbb4d34288d7e54be431e0d3772b144177ad70c33076be31027f900812046\": container with ID starting with 9cdbbb4d34288d7e54be431e0d3772b144177ad70c33076be31027f900812046 not found: ID does not exist" containerID="9cdbbb4d34288d7e54be431e0d3772b144177ad70c33076be31027f900812046" Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.693184 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cdbbb4d34288d7e54be431e0d3772b144177ad70c33076be31027f900812046"} err="failed to get container status \"9cdbbb4d34288d7e54be431e0d3772b144177ad70c33076be31027f900812046\": rpc error: code = NotFound desc = could not find container \"9cdbbb4d34288d7e54be431e0d3772b144177ad70c33076be31027f900812046\": container with ID starting with 9cdbbb4d34288d7e54be431e0d3772b144177ad70c33076be31027f900812046 not found: ID does not exist" Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.693223 4814 scope.go:117] "RemoveContainer" containerID="61821c5666cd74e8a1ef8c012e45d23be1cb6eaed2afc33cf8b6d161824d6fc7" Feb 27 17:28:28 crc kubenswrapper[4814]: E0227 17:28:28.693838 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61821c5666cd74e8a1ef8c012e45d23be1cb6eaed2afc33cf8b6d161824d6fc7\": container with ID starting with 61821c5666cd74e8a1ef8c012e45d23be1cb6eaed2afc33cf8b6d161824d6fc7 not found: ID does not exist" containerID="61821c5666cd74e8a1ef8c012e45d23be1cb6eaed2afc33cf8b6d161824d6fc7" Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.693878 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61821c5666cd74e8a1ef8c012e45d23be1cb6eaed2afc33cf8b6d161824d6fc7"} err="failed to get container status \"61821c5666cd74e8a1ef8c012e45d23be1cb6eaed2afc33cf8b6d161824d6fc7\": rpc error: code = NotFound desc = could not find container \"61821c5666cd74e8a1ef8c012e45d23be1cb6eaed2afc33cf8b6d161824d6fc7\": container with ID starting with 61821c5666cd74e8a1ef8c012e45d23be1cb6eaed2afc33cf8b6d161824d6fc7 not found: ID does not exist" Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.693903 4814 scope.go:117] "RemoveContainer" containerID="95864b6707e4e5b1d402ade41538fbc4d4bf54f88e5aa4792d270cc5dcf91992" Feb 27 17:28:28 crc kubenswrapper[4814]: E0227 17:28:28.694171 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95864b6707e4e5b1d402ade41538fbc4d4bf54f88e5aa4792d270cc5dcf91992\": container with ID starting with 95864b6707e4e5b1d402ade41538fbc4d4bf54f88e5aa4792d270cc5dcf91992 not found: ID does not exist" containerID="95864b6707e4e5b1d402ade41538fbc4d4bf54f88e5aa4792d270cc5dcf91992" Feb 27 17:28:28 crc kubenswrapper[4814]: I0227 17:28:28.694211 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95864b6707e4e5b1d402ade41538fbc4d4bf54f88e5aa4792d270cc5dcf91992"} err="failed to get container status \"95864b6707e4e5b1d402ade41538fbc4d4bf54f88e5aa4792d270cc5dcf91992\": rpc error: code = NotFound desc = could not find container \"95864b6707e4e5b1d402ade41538fbc4d4bf54f88e5aa4792d270cc5dcf91992\": container with ID starting with 95864b6707e4e5b1d402ade41538fbc4d4bf54f88e5aa4792d270cc5dcf91992 not found: ID does not exist" Feb 27 17:28:30 crc kubenswrapper[4814]: I0227 17:28:30.509775 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae394f60-eb2c-4b59-ba67-bb7a3e34ef10" path="/var/lib/kubelet/pods/ae394f60-eb2c-4b59-ba67-bb7a3e34ef10/volumes" Feb 27 17:28:52 crc kubenswrapper[4814]: I0227 17:28:52.903219 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:28:52 crc kubenswrapper[4814]: I0227 17:28:52.905214 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:28:52 crc kubenswrapper[4814]: I0227 17:28:52.905333 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 17:28:52 crc kubenswrapper[4814]: I0227 17:28:52.906655 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3"} pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:28:52 crc kubenswrapper[4814]: I0227 17:28:52.906734 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" containerID="cri-o://e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" gracePeriod=600 Feb 27 17:28:53 crc kubenswrapper[4814]: E0227 17:28:53.038962 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:28:53 crc kubenswrapper[4814]: I0227 17:28:53.858981 4814 generic.go:334] "Generic (PLEG): container finished" podID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" exitCode=0 Feb 27 17:28:53 crc kubenswrapper[4814]: I0227 17:28:53.859088 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerDied","Data":"e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3"} Feb 27 17:28:53 crc kubenswrapper[4814]: I0227 17:28:53.859503 4814 scope.go:117] "RemoveContainer" containerID="daa4ecfa76d33456eca88bbfe3cf3adb033c01a242837064c2ae775d74b120a7" Feb 27 17:28:53 crc kubenswrapper[4814]: I0227 17:28:53.860618 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:28:53 crc kubenswrapper[4814]: E0227 17:28:53.861286 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:29:04 crc kubenswrapper[4814]: I0227 17:29:04.488561 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:29:04 crc kubenswrapper[4814]: E0227 17:29:04.489910 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:29:18 crc kubenswrapper[4814]: I0227 17:29:18.504503 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:29:18 crc kubenswrapper[4814]: E0227 17:29:18.506218 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:29:30 crc kubenswrapper[4814]: E0227 17:29:30.019219 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Feb 27 17:29:33 crc kubenswrapper[4814]: I0227 17:29:33.488328 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:29:33 crc kubenswrapper[4814]: E0227 17:29:33.489292 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:29:45 crc kubenswrapper[4814]: I0227 17:29:45.489019 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:29:45 crc kubenswrapper[4814]: E0227 17:29:45.490492 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:29:59 crc kubenswrapper[4814]: I0227 17:29:59.489187 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:29:59 crc kubenswrapper[4814]: E0227 17:29:59.490911 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.180306 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536890-r9hrf"] Feb 27 17:30:00 crc kubenswrapper[4814]: E0227 17:30:00.180886 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae394f60-eb2c-4b59-ba67-bb7a3e34ef10" containerName="extract-utilities" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.180911 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae394f60-eb2c-4b59-ba67-bb7a3e34ef10" containerName="extract-utilities" Feb 27 17:30:00 crc kubenswrapper[4814]: E0227 17:30:00.180943 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae394f60-eb2c-4b59-ba67-bb7a3e34ef10" containerName="extract-content" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.180957 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae394f60-eb2c-4b59-ba67-bb7a3e34ef10" containerName="extract-content" Feb 27 17:30:00 crc kubenswrapper[4814]: E0227 17:30:00.180998 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae394f60-eb2c-4b59-ba67-bb7a3e34ef10" containerName="registry-server" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.181011 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae394f60-eb2c-4b59-ba67-bb7a3e34ef10" containerName="registry-server" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.181395 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae394f60-eb2c-4b59-ba67-bb7a3e34ef10" containerName="registry-server" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.182452 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536890-r9hrf" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.185512 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.186602 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.191922 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.200472 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b"] Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.202343 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.204804 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.205456 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.217741 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536890-r9hrf"] Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.227960 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b"] Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.233840 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwtgg\" (UniqueName: \"kubernetes.io/projected/90efa58d-0276-4e00-af07-2a098c5b06a9-kube-api-access-wwtgg\") pod \"auto-csr-approver-29536890-r9hrf\" (UID: \"90efa58d-0276-4e00-af07-2a098c5b06a9\") " pod="openshift-infra/auto-csr-approver-29536890-r9hrf" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.233929 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-config-volume\") pod \"collect-profiles-29536890-86x6b\" (UID: \"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.233982 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-secret-volume\") pod \"collect-profiles-29536890-86x6b\" (UID: \"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.234101 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-668gq\" (UniqueName: \"kubernetes.io/projected/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-kube-api-access-668gq\") pod \"collect-profiles-29536890-86x6b\" (UID: \"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.336393 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwtgg\" (UniqueName: \"kubernetes.io/projected/90efa58d-0276-4e00-af07-2a098c5b06a9-kube-api-access-wwtgg\") pod \"auto-csr-approver-29536890-r9hrf\" (UID: \"90efa58d-0276-4e00-af07-2a098c5b06a9\") " pod="openshift-infra/auto-csr-approver-29536890-r9hrf" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.336482 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-config-volume\") pod \"collect-profiles-29536890-86x6b\" (UID: \"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.336522 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-secret-volume\") pod \"collect-profiles-29536890-86x6b\" (UID: \"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.336617 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-668gq\" (UniqueName: \"kubernetes.io/projected/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-kube-api-access-668gq\") pod \"collect-profiles-29536890-86x6b\" (UID: \"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.338108 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-config-volume\") pod \"collect-profiles-29536890-86x6b\" (UID: \"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.357644 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-secret-volume\") pod \"collect-profiles-29536890-86x6b\" (UID: \"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.357868 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-668gq\" (UniqueName: \"kubernetes.io/projected/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-kube-api-access-668gq\") pod \"collect-profiles-29536890-86x6b\" (UID: \"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.362279 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwtgg\" (UniqueName: \"kubernetes.io/projected/90efa58d-0276-4e00-af07-2a098c5b06a9-kube-api-access-wwtgg\") pod \"auto-csr-approver-29536890-r9hrf\" (UID: \"90efa58d-0276-4e00-af07-2a098c5b06a9\") " pod="openshift-infra/auto-csr-approver-29536890-r9hrf" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.511933 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536890-r9hrf" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.538202 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b" Feb 27 17:30:00 crc kubenswrapper[4814]: I0227 17:30:00.913899 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536890-r9hrf"] Feb 27 17:30:01 crc kubenswrapper[4814]: I0227 17:30:01.056052 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b"] Feb 27 17:30:01 crc kubenswrapper[4814]: W0227 17:30:01.063865 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0d6d1da_289d_4b32_9f3b_ed0a244fcd77.slice/crio-b4922918231ab3dcf65ff9aa39024e845ee9d664e07d087b8dd694a2c16390e7 WatchSource:0}: Error finding container b4922918231ab3dcf65ff9aa39024e845ee9d664e07d087b8dd694a2c16390e7: Status 404 returned error can't find the container with id b4922918231ab3dcf65ff9aa39024e845ee9d664e07d087b8dd694a2c16390e7 Feb 27 17:30:01 crc kubenswrapper[4814]: I0227 17:30:01.743682 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536890-r9hrf" event={"ID":"90efa58d-0276-4e00-af07-2a098c5b06a9","Type":"ContainerStarted","Data":"9e14906b0ea8b5cec6d9e408c27072defb9031185dfc4975f0c0454c4594c63e"} Feb 27 17:30:01 crc kubenswrapper[4814]: I0227 17:30:01.746482 4814 generic.go:334] "Generic (PLEG): container finished" podID="d0d6d1da-289d-4b32-9f3b-ed0a244fcd77" containerID="5038c628fbaa55a295dbb63b2268554f9fccdc9591aebd710757d67e5c02eab3" exitCode=0 Feb 27 17:30:01 crc kubenswrapper[4814]: I0227 17:30:01.746518 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b" event={"ID":"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77","Type":"ContainerDied","Data":"5038c628fbaa55a295dbb63b2268554f9fccdc9591aebd710757d67e5c02eab3"} Feb 27 17:30:01 crc kubenswrapper[4814]: I0227 17:30:01.746541 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b" event={"ID":"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77","Type":"ContainerStarted","Data":"b4922918231ab3dcf65ff9aa39024e845ee9d664e07d087b8dd694a2c16390e7"} Feb 27 17:30:02 crc kubenswrapper[4814]: I0227 17:30:02.762351 4814 generic.go:334] "Generic (PLEG): container finished" podID="90efa58d-0276-4e00-af07-2a098c5b06a9" containerID="26dd361bfa4b030b453619c08d34ebda7923226905b5b66c0ec43566e76cd32e" exitCode=0 Feb 27 17:30:02 crc kubenswrapper[4814]: I0227 17:30:02.762871 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536890-r9hrf" event={"ID":"90efa58d-0276-4e00-af07-2a098c5b06a9","Type":"ContainerDied","Data":"26dd361bfa4b030b453619c08d34ebda7923226905b5b66c0ec43566e76cd32e"} Feb 27 17:30:03 crc kubenswrapper[4814]: I0227 17:30:03.278472 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b" Feb 27 17:30:03 crc kubenswrapper[4814]: I0227 17:30:03.422363 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-668gq\" (UniqueName: \"kubernetes.io/projected/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-kube-api-access-668gq\") pod \"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77\" (UID: \"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77\") " Feb 27 17:30:03 crc kubenswrapper[4814]: I0227 17:30:03.422444 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-config-volume\") pod \"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77\" (UID: \"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77\") " Feb 27 17:30:03 crc kubenswrapper[4814]: I0227 17:30:03.422575 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-secret-volume\") pod \"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77\" (UID: \"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77\") " Feb 27 17:30:03 crc kubenswrapper[4814]: I0227 17:30:03.425729 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-config-volume" (OuterVolumeSpecName: "config-volume") pod "d0d6d1da-289d-4b32-9f3b-ed0a244fcd77" (UID: "d0d6d1da-289d-4b32-9f3b-ed0a244fcd77"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:30:03 crc kubenswrapper[4814]: I0227 17:30:03.433676 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d0d6d1da-289d-4b32-9f3b-ed0a244fcd77" (UID: "d0d6d1da-289d-4b32-9f3b-ed0a244fcd77"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:30:03 crc kubenswrapper[4814]: I0227 17:30:03.433978 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-kube-api-access-668gq" (OuterVolumeSpecName: "kube-api-access-668gq") pod "d0d6d1da-289d-4b32-9f3b-ed0a244fcd77" (UID: "d0d6d1da-289d-4b32-9f3b-ed0a244fcd77"). InnerVolumeSpecName "kube-api-access-668gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:30:03 crc kubenswrapper[4814]: I0227 17:30:03.527319 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-668gq\" (UniqueName: \"kubernetes.io/projected/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-kube-api-access-668gq\") on node \"crc\" DevicePath \"\"" Feb 27 17:30:03 crc kubenswrapper[4814]: I0227 17:30:03.527428 4814 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 17:30:03 crc kubenswrapper[4814]: I0227 17:30:03.527501 4814 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0d6d1da-289d-4b32-9f3b-ed0a244fcd77-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 17:30:03 crc kubenswrapper[4814]: I0227 17:30:03.789598 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b" Feb 27 17:30:03 crc kubenswrapper[4814]: I0227 17:30:03.791572 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-86x6b" event={"ID":"d0d6d1da-289d-4b32-9f3b-ed0a244fcd77","Type":"ContainerDied","Data":"b4922918231ab3dcf65ff9aa39024e845ee9d664e07d087b8dd694a2c16390e7"} Feb 27 17:30:03 crc kubenswrapper[4814]: I0227 17:30:03.791690 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4922918231ab3dcf65ff9aa39024e845ee9d664e07d087b8dd694a2c16390e7" Feb 27 17:30:04 crc kubenswrapper[4814]: I0227 17:30:04.377670 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536890-r9hrf" Feb 27 17:30:04 crc kubenswrapper[4814]: I0227 17:30:04.392456 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq"] Feb 27 17:30:04 crc kubenswrapper[4814]: I0227 17:30:04.404181 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536845-zzkrq"] Feb 27 17:30:04 crc kubenswrapper[4814]: I0227 17:30:04.457862 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwtgg\" (UniqueName: \"kubernetes.io/projected/90efa58d-0276-4e00-af07-2a098c5b06a9-kube-api-access-wwtgg\") pod \"90efa58d-0276-4e00-af07-2a098c5b06a9\" (UID: \"90efa58d-0276-4e00-af07-2a098c5b06a9\") " Feb 27 17:30:04 crc kubenswrapper[4814]: I0227 17:30:04.463195 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90efa58d-0276-4e00-af07-2a098c5b06a9-kube-api-access-wwtgg" (OuterVolumeSpecName: "kube-api-access-wwtgg") pod "90efa58d-0276-4e00-af07-2a098c5b06a9" (UID: "90efa58d-0276-4e00-af07-2a098c5b06a9"). InnerVolumeSpecName "kube-api-access-wwtgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:30:04 crc kubenswrapper[4814]: I0227 17:30:04.506666 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa51609e-188a-4ded-962b-9f9641404afc" path="/var/lib/kubelet/pods/aa51609e-188a-4ded-962b-9f9641404afc/volumes" Feb 27 17:30:04 crc kubenswrapper[4814]: I0227 17:30:04.559755 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwtgg\" (UniqueName: \"kubernetes.io/projected/90efa58d-0276-4e00-af07-2a098c5b06a9-kube-api-access-wwtgg\") on node \"crc\" DevicePath \"\"" Feb 27 17:30:04 crc kubenswrapper[4814]: I0227 17:30:04.806319 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536890-r9hrf" event={"ID":"90efa58d-0276-4e00-af07-2a098c5b06a9","Type":"ContainerDied","Data":"9e14906b0ea8b5cec6d9e408c27072defb9031185dfc4975f0c0454c4594c63e"} Feb 27 17:30:04 crc kubenswrapper[4814]: I0227 17:30:04.806371 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e14906b0ea8b5cec6d9e408c27072defb9031185dfc4975f0c0454c4594c63e" Feb 27 17:30:04 crc kubenswrapper[4814]: I0227 17:30:04.806450 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536890-r9hrf" Feb 27 17:30:05 crc kubenswrapper[4814]: I0227 17:30:05.472923 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536884-lb8rx"] Feb 27 17:30:05 crc kubenswrapper[4814]: I0227 17:30:05.487335 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536884-lb8rx"] Feb 27 17:30:06 crc kubenswrapper[4814]: I0227 17:30:06.510323 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3596843e-9ae6-45eb-b1c9-e82215ced5e6" path="/var/lib/kubelet/pods/3596843e-9ae6-45eb-b1c9-e82215ced5e6/volumes" Feb 27 17:30:07 crc kubenswrapper[4814]: I0227 17:30:07.832211 4814 scope.go:117] "RemoveContainer" containerID="32160fe46f2873c0b65fc5338bb1f9d69c3d5960d23dcb8b595f7c7112647eaf" Feb 27 17:30:07 crc kubenswrapper[4814]: I0227 17:30:07.953468 4814 scope.go:117] "RemoveContainer" containerID="a7fbeff9a2f69b9ac82fe3541fa4c702b7fc9c91b28f6827d4bef735d3c13471" Feb 27 17:30:11 crc kubenswrapper[4814]: I0227 17:30:11.487503 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:30:11 crc kubenswrapper[4814]: E0227 17:30:11.488545 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:30:13 crc kubenswrapper[4814]: I0227 17:30:13.928651 4814 generic.go:334] "Generic (PLEG): container finished" podID="0e6b3bb7-c25b-491c-a582-a73eb2c15d85" containerID="f0d84c7eceeb93d66fcb4e5813c0721df388bd53a4f150d8e4a59287d656e14e" exitCode=0 Feb 27 17:30:13 crc kubenswrapper[4814]: I0227 17:30:13.928799 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6tx9/must-gather-mghpl" event={"ID":"0e6b3bb7-c25b-491c-a582-a73eb2c15d85","Type":"ContainerDied","Data":"f0d84c7eceeb93d66fcb4e5813c0721df388bd53a4f150d8e4a59287d656e14e"} Feb 27 17:30:13 crc kubenswrapper[4814]: I0227 17:30:13.930290 4814 scope.go:117] "RemoveContainer" containerID="f0d84c7eceeb93d66fcb4e5813c0721df388bd53a4f150d8e4a59287d656e14e" Feb 27 17:30:14 crc kubenswrapper[4814]: I0227 17:30:14.424830 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t6tx9_must-gather-mghpl_0e6b3bb7-c25b-491c-a582-a73eb2c15d85/gather/0.log" Feb 27 17:30:22 crc kubenswrapper[4814]: I0227 17:30:22.241421 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t6tx9/must-gather-mghpl"] Feb 27 17:30:22 crc kubenswrapper[4814]: I0227 17:30:22.243308 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-t6tx9/must-gather-mghpl" podUID="0e6b3bb7-c25b-491c-a582-a73eb2c15d85" containerName="copy" containerID="cri-o://9d1bf312283009abb0f306d7c21ee78c73e613df948a46d0d87df7ed541b216f" gracePeriod=2 Feb 27 17:30:22 crc kubenswrapper[4814]: I0227 17:30:22.249947 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t6tx9/must-gather-mghpl"] Feb 27 17:30:22 crc kubenswrapper[4814]: I0227 17:30:22.712269 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t6tx9_must-gather-mghpl_0e6b3bb7-c25b-491c-a582-a73eb2c15d85/copy/0.log" Feb 27 17:30:22 crc kubenswrapper[4814]: I0227 17:30:22.713169 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6tx9/must-gather-mghpl" Feb 27 17:30:22 crc kubenswrapper[4814]: I0227 17:30:22.825339 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0e6b3bb7-c25b-491c-a582-a73eb2c15d85-must-gather-output\") pod \"0e6b3bb7-c25b-491c-a582-a73eb2c15d85\" (UID: \"0e6b3bb7-c25b-491c-a582-a73eb2c15d85\") " Feb 27 17:30:22 crc kubenswrapper[4814]: I0227 17:30:22.825520 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgstx\" (UniqueName: \"kubernetes.io/projected/0e6b3bb7-c25b-491c-a582-a73eb2c15d85-kube-api-access-rgstx\") pod \"0e6b3bb7-c25b-491c-a582-a73eb2c15d85\" (UID: \"0e6b3bb7-c25b-491c-a582-a73eb2c15d85\") " Feb 27 17:30:22 crc kubenswrapper[4814]: I0227 17:30:22.831671 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e6b3bb7-c25b-491c-a582-a73eb2c15d85-kube-api-access-rgstx" (OuterVolumeSpecName: "kube-api-access-rgstx") pod "0e6b3bb7-c25b-491c-a582-a73eb2c15d85" (UID: "0e6b3bb7-c25b-491c-a582-a73eb2c15d85"). InnerVolumeSpecName "kube-api-access-rgstx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:30:22 crc kubenswrapper[4814]: I0227 17:30:22.928073 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgstx\" (UniqueName: \"kubernetes.io/projected/0e6b3bb7-c25b-491c-a582-a73eb2c15d85-kube-api-access-rgstx\") on node \"crc\" DevicePath \"\"" Feb 27 17:30:22 crc kubenswrapper[4814]: I0227 17:30:22.982795 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e6b3bb7-c25b-491c-a582-a73eb2c15d85-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "0e6b3bb7-c25b-491c-a582-a73eb2c15d85" (UID: "0e6b3bb7-c25b-491c-a582-a73eb2c15d85"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:30:23 crc kubenswrapper[4814]: I0227 17:30:23.030322 4814 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0e6b3bb7-c25b-491c-a582-a73eb2c15d85-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 27 17:30:23 crc kubenswrapper[4814]: I0227 17:30:23.044397 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t6tx9_must-gather-mghpl_0e6b3bb7-c25b-491c-a582-a73eb2c15d85/copy/0.log" Feb 27 17:30:23 crc kubenswrapper[4814]: I0227 17:30:23.045146 4814 generic.go:334] "Generic (PLEG): container finished" podID="0e6b3bb7-c25b-491c-a582-a73eb2c15d85" containerID="9d1bf312283009abb0f306d7c21ee78c73e613df948a46d0d87df7ed541b216f" exitCode=143 Feb 27 17:30:23 crc kubenswrapper[4814]: I0227 17:30:23.045299 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6tx9/must-gather-mghpl" Feb 27 17:30:23 crc kubenswrapper[4814]: I0227 17:30:23.045371 4814 scope.go:117] "RemoveContainer" containerID="9d1bf312283009abb0f306d7c21ee78c73e613df948a46d0d87df7ed541b216f" Feb 27 17:30:23 crc kubenswrapper[4814]: I0227 17:30:23.080535 4814 scope.go:117] "RemoveContainer" containerID="f0d84c7eceeb93d66fcb4e5813c0721df388bd53a4f150d8e4a59287d656e14e" Feb 27 17:30:23 crc kubenswrapper[4814]: I0227 17:30:23.207041 4814 scope.go:117] "RemoveContainer" containerID="9d1bf312283009abb0f306d7c21ee78c73e613df948a46d0d87df7ed541b216f" Feb 27 17:30:23 crc kubenswrapper[4814]: E0227 17:30:23.207625 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d1bf312283009abb0f306d7c21ee78c73e613df948a46d0d87df7ed541b216f\": container with ID starting with 9d1bf312283009abb0f306d7c21ee78c73e613df948a46d0d87df7ed541b216f not found: ID does not exist" containerID="9d1bf312283009abb0f306d7c21ee78c73e613df948a46d0d87df7ed541b216f" Feb 27 17:30:23 crc kubenswrapper[4814]: I0227 17:30:23.207687 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d1bf312283009abb0f306d7c21ee78c73e613df948a46d0d87df7ed541b216f"} err="failed to get container status \"9d1bf312283009abb0f306d7c21ee78c73e613df948a46d0d87df7ed541b216f\": rpc error: code = NotFound desc = could not find container \"9d1bf312283009abb0f306d7c21ee78c73e613df948a46d0d87df7ed541b216f\": container with ID starting with 9d1bf312283009abb0f306d7c21ee78c73e613df948a46d0d87df7ed541b216f not found: ID does not exist" Feb 27 17:30:23 crc kubenswrapper[4814]: I0227 17:30:23.207722 4814 scope.go:117] "RemoveContainer" containerID="f0d84c7eceeb93d66fcb4e5813c0721df388bd53a4f150d8e4a59287d656e14e" Feb 27 17:30:23 crc kubenswrapper[4814]: E0227 17:30:23.208080 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0d84c7eceeb93d66fcb4e5813c0721df388bd53a4f150d8e4a59287d656e14e\": container with ID starting with f0d84c7eceeb93d66fcb4e5813c0721df388bd53a4f150d8e4a59287d656e14e not found: ID does not exist" containerID="f0d84c7eceeb93d66fcb4e5813c0721df388bd53a4f150d8e4a59287d656e14e" Feb 27 17:30:23 crc kubenswrapper[4814]: I0227 17:30:23.208123 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0d84c7eceeb93d66fcb4e5813c0721df388bd53a4f150d8e4a59287d656e14e"} err="failed to get container status \"f0d84c7eceeb93d66fcb4e5813c0721df388bd53a4f150d8e4a59287d656e14e\": rpc error: code = NotFound desc = could not find container \"f0d84c7eceeb93d66fcb4e5813c0721df388bd53a4f150d8e4a59287d656e14e\": container with ID starting with f0d84c7eceeb93d66fcb4e5813c0721df388bd53a4f150d8e4a59287d656e14e not found: ID does not exist" Feb 27 17:30:23 crc kubenswrapper[4814]: I0227 17:30:23.487772 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:30:23 crc kubenswrapper[4814]: E0227 17:30:23.488402 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:30:24 crc kubenswrapper[4814]: I0227 17:30:24.499311 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e6b3bb7-c25b-491c-a582-a73eb2c15d85" path="/var/lib/kubelet/pods/0e6b3bb7-c25b-491c-a582-a73eb2c15d85/volumes" Feb 27 17:30:34 crc kubenswrapper[4814]: I0227 17:30:34.489653 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:30:34 crc kubenswrapper[4814]: E0227 17:30:34.490761 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:30:48 crc kubenswrapper[4814]: I0227 17:30:48.496383 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:30:48 crc kubenswrapper[4814]: E0227 17:30:48.497546 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:30:59 crc kubenswrapper[4814]: I0227 17:30:59.488355 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:30:59 crc kubenswrapper[4814]: E0227 17:30:59.489454 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:31:08 crc kubenswrapper[4814]: I0227 17:31:08.018121 4814 scope.go:117] "RemoveContainer" containerID="fdb18e92d8fbcf13ac997fc037379638ad73bc61cbd4c6d671c4b3a187d1f1f1" Feb 27 17:31:12 crc kubenswrapper[4814]: I0227 17:31:12.488917 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:31:12 crc kubenswrapper[4814]: E0227 17:31:12.490936 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:31:27 crc kubenswrapper[4814]: I0227 17:31:27.489388 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:31:27 crc kubenswrapper[4814]: E0227 17:31:27.491059 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:31:40 crc kubenswrapper[4814]: I0227 17:31:40.488053 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:31:40 crc kubenswrapper[4814]: E0227 17:31:40.489037 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.413993 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gpnkz"] Feb 27 17:31:45 crc kubenswrapper[4814]: E0227 17:31:45.415715 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e6b3bb7-c25b-491c-a582-a73eb2c15d85" containerName="copy" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.415742 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e6b3bb7-c25b-491c-a582-a73eb2c15d85" containerName="copy" Feb 27 17:31:45 crc kubenswrapper[4814]: E0227 17:31:45.415777 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90efa58d-0276-4e00-af07-2a098c5b06a9" containerName="oc" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.415789 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="90efa58d-0276-4e00-af07-2a098c5b06a9" containerName="oc" Feb 27 17:31:45 crc kubenswrapper[4814]: E0227 17:31:45.415833 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e6b3bb7-c25b-491c-a582-a73eb2c15d85" containerName="gather" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.415847 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e6b3bb7-c25b-491c-a582-a73eb2c15d85" containerName="gather" Feb 27 17:31:45 crc kubenswrapper[4814]: E0227 17:31:45.415872 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d6d1da-289d-4b32-9f3b-ed0a244fcd77" containerName="collect-profiles" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.415885 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d6d1da-289d-4b32-9f3b-ed0a244fcd77" containerName="collect-profiles" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.416216 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e6b3bb7-c25b-491c-a582-a73eb2c15d85" containerName="copy" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.416280 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e6b3bb7-c25b-491c-a582-a73eb2c15d85" containerName="gather" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.416312 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0d6d1da-289d-4b32-9f3b-ed0a244fcd77" containerName="collect-profiles" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.416339 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="90efa58d-0276-4e00-af07-2a098c5b06a9" containerName="oc" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.418887 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.436767 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gpnkz"] Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.442209 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c016a833-d123-4840-ac77-72be85b6cab5-utilities\") pod \"redhat-operators-gpnkz\" (UID: \"c016a833-d123-4840-ac77-72be85b6cab5\") " pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.442673 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxnxz\" (UniqueName: \"kubernetes.io/projected/c016a833-d123-4840-ac77-72be85b6cab5-kube-api-access-lxnxz\") pod \"redhat-operators-gpnkz\" (UID: \"c016a833-d123-4840-ac77-72be85b6cab5\") " pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.442722 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c016a833-d123-4840-ac77-72be85b6cab5-catalog-content\") pod \"redhat-operators-gpnkz\" (UID: \"c016a833-d123-4840-ac77-72be85b6cab5\") " pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.545443 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxnxz\" (UniqueName: \"kubernetes.io/projected/c016a833-d123-4840-ac77-72be85b6cab5-kube-api-access-lxnxz\") pod \"redhat-operators-gpnkz\" (UID: \"c016a833-d123-4840-ac77-72be85b6cab5\") " pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.545524 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c016a833-d123-4840-ac77-72be85b6cab5-catalog-content\") pod \"redhat-operators-gpnkz\" (UID: \"c016a833-d123-4840-ac77-72be85b6cab5\") " pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.545747 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c016a833-d123-4840-ac77-72be85b6cab5-utilities\") pod \"redhat-operators-gpnkz\" (UID: \"c016a833-d123-4840-ac77-72be85b6cab5\") " pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.546504 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c016a833-d123-4840-ac77-72be85b6cab5-utilities\") pod \"redhat-operators-gpnkz\" (UID: \"c016a833-d123-4840-ac77-72be85b6cab5\") " pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.546981 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c016a833-d123-4840-ac77-72be85b6cab5-catalog-content\") pod \"redhat-operators-gpnkz\" (UID: \"c016a833-d123-4840-ac77-72be85b6cab5\") " pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.570807 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxnxz\" (UniqueName: \"kubernetes.io/projected/c016a833-d123-4840-ac77-72be85b6cab5-kube-api-access-lxnxz\") pod \"redhat-operators-gpnkz\" (UID: \"c016a833-d123-4840-ac77-72be85b6cab5\") " pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:31:45 crc kubenswrapper[4814]: I0227 17:31:45.750318 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:31:46 crc kubenswrapper[4814]: I0227 17:31:46.223143 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gpnkz"] Feb 27 17:31:47 crc kubenswrapper[4814]: I0227 17:31:47.168384 4814 generic.go:334] "Generic (PLEG): container finished" podID="c016a833-d123-4840-ac77-72be85b6cab5" containerID="2b728196e7a85b4ca25577045b89a5e205e7c858fd2982ac5a1138da36534263" exitCode=0 Feb 27 17:31:47 crc kubenswrapper[4814]: I0227 17:31:47.168469 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gpnkz" event={"ID":"c016a833-d123-4840-ac77-72be85b6cab5","Type":"ContainerDied","Data":"2b728196e7a85b4ca25577045b89a5e205e7c858fd2982ac5a1138da36534263"} Feb 27 17:31:47 crc kubenswrapper[4814]: I0227 17:31:47.168938 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gpnkz" event={"ID":"c016a833-d123-4840-ac77-72be85b6cab5","Type":"ContainerStarted","Data":"2a4e80e975a710f4230c976a069ec46d537ba07d1d5756fb97c99874fd5fe8bf"} Feb 27 17:31:47 crc kubenswrapper[4814]: I0227 17:31:47.171914 4814 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:31:49 crc kubenswrapper[4814]: I0227 17:31:49.196607 4814 generic.go:334] "Generic (PLEG): container finished" podID="c016a833-d123-4840-ac77-72be85b6cab5" containerID="bc7f04db634bfd6c19ccbb2716748bb1b3c53b18fe0735f1c209d002a175aacd" exitCode=0 Feb 27 17:31:49 crc kubenswrapper[4814]: I0227 17:31:49.196737 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gpnkz" event={"ID":"c016a833-d123-4840-ac77-72be85b6cab5","Type":"ContainerDied","Data":"bc7f04db634bfd6c19ccbb2716748bb1b3c53b18fe0735f1c209d002a175aacd"} Feb 27 17:31:50 crc kubenswrapper[4814]: I0227 17:31:50.216597 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gpnkz" event={"ID":"c016a833-d123-4840-ac77-72be85b6cab5","Type":"ContainerStarted","Data":"002186015d5fc630f91891c18f75d0e9c07e36bf888e309932081368fb45a05c"} Feb 27 17:31:50 crc kubenswrapper[4814]: I0227 17:31:50.253662 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gpnkz" podStartSLOduration=2.781260847 podStartE2EDuration="5.253626391s" podCreationTimestamp="2026-02-27 17:31:45 +0000 UTC" firstStartedPulling="2026-02-27 17:31:47.171454929 +0000 UTC m=+4119.624079799" lastFinishedPulling="2026-02-27 17:31:49.643820513 +0000 UTC m=+4122.096445343" observedRunningTime="2026-02-27 17:31:50.239622786 +0000 UTC m=+4122.692247616" watchObservedRunningTime="2026-02-27 17:31:50.253626391 +0000 UTC m=+4122.706251251" Feb 27 17:31:53 crc kubenswrapper[4814]: I0227 17:31:53.488391 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:31:53 crc kubenswrapper[4814]: E0227 17:31:53.489407 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:31:55 crc kubenswrapper[4814]: I0227 17:31:55.751706 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:31:55 crc kubenswrapper[4814]: I0227 17:31:55.753054 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:31:56 crc kubenswrapper[4814]: I0227 17:31:56.836652 4814 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gpnkz" podUID="c016a833-d123-4840-ac77-72be85b6cab5" containerName="registry-server" probeResult="failure" output=< Feb 27 17:31:56 crc kubenswrapper[4814]: timeout: failed to connect service ":50051" within 1s Feb 27 17:31:56 crc kubenswrapper[4814]: > Feb 27 17:32:00 crc kubenswrapper[4814]: I0227 17:32:00.177474 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536892-6ktg6"] Feb 27 17:32:00 crc kubenswrapper[4814]: I0227 17:32:00.180387 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536892-6ktg6" Feb 27 17:32:00 crc kubenswrapper[4814]: I0227 17:32:00.184320 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:32:00 crc kubenswrapper[4814]: I0227 17:32:00.184524 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:32:00 crc kubenswrapper[4814]: I0227 17:32:00.185148 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:32:00 crc kubenswrapper[4814]: I0227 17:32:00.193484 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536892-6ktg6"] Feb 27 17:32:00 crc kubenswrapper[4814]: I0227 17:32:00.247233 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r4vd\" (UniqueName: \"kubernetes.io/projected/2e9b3520-7570-4c37-8fec-58b0de0af5a9-kube-api-access-5r4vd\") pod \"auto-csr-approver-29536892-6ktg6\" (UID: \"2e9b3520-7570-4c37-8fec-58b0de0af5a9\") " pod="openshift-infra/auto-csr-approver-29536892-6ktg6" Feb 27 17:32:00 crc kubenswrapper[4814]: I0227 17:32:00.349742 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r4vd\" (UniqueName: \"kubernetes.io/projected/2e9b3520-7570-4c37-8fec-58b0de0af5a9-kube-api-access-5r4vd\") pod \"auto-csr-approver-29536892-6ktg6\" (UID: \"2e9b3520-7570-4c37-8fec-58b0de0af5a9\") " pod="openshift-infra/auto-csr-approver-29536892-6ktg6" Feb 27 17:32:00 crc kubenswrapper[4814]: I0227 17:32:00.374278 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r4vd\" (UniqueName: \"kubernetes.io/projected/2e9b3520-7570-4c37-8fec-58b0de0af5a9-kube-api-access-5r4vd\") pod \"auto-csr-approver-29536892-6ktg6\" (UID: \"2e9b3520-7570-4c37-8fec-58b0de0af5a9\") " pod="openshift-infra/auto-csr-approver-29536892-6ktg6" Feb 27 17:32:00 crc kubenswrapper[4814]: I0227 17:32:00.518976 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536892-6ktg6" Feb 27 17:32:01 crc kubenswrapper[4814]: I0227 17:32:01.038509 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536892-6ktg6"] Feb 27 17:32:01 crc kubenswrapper[4814]: W0227 17:32:01.589979 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e9b3520_7570_4c37_8fec_58b0de0af5a9.slice/crio-388afd28e6592119d8d3fd80b3b0c7fa0d1f9d0e0bb0bf2b8948007374dfc959 WatchSource:0}: Error finding container 388afd28e6592119d8d3fd80b3b0c7fa0d1f9d0e0bb0bf2b8948007374dfc959: Status 404 returned error can't find the container with id 388afd28e6592119d8d3fd80b3b0c7fa0d1f9d0e0bb0bf2b8948007374dfc959 Feb 27 17:32:02 crc kubenswrapper[4814]: I0227 17:32:02.384517 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536892-6ktg6" event={"ID":"2e9b3520-7570-4c37-8fec-58b0de0af5a9","Type":"ContainerStarted","Data":"388afd28e6592119d8d3fd80b3b0c7fa0d1f9d0e0bb0bf2b8948007374dfc959"} Feb 27 17:32:03 crc kubenswrapper[4814]: I0227 17:32:03.400827 4814 generic.go:334] "Generic (PLEG): container finished" podID="2e9b3520-7570-4c37-8fec-58b0de0af5a9" containerID="6ed906f1596af25da928a75c1dcce55b92563606220e928058dc6ee2b9afa179" exitCode=0 Feb 27 17:32:03 crc kubenswrapper[4814]: I0227 17:32:03.400970 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536892-6ktg6" event={"ID":"2e9b3520-7570-4c37-8fec-58b0de0af5a9","Type":"ContainerDied","Data":"6ed906f1596af25da928a75c1dcce55b92563606220e928058dc6ee2b9afa179"} Feb 27 17:32:05 crc kubenswrapper[4814]: I0227 17:32:05.442721 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536892-6ktg6" event={"ID":"2e9b3520-7570-4c37-8fec-58b0de0af5a9","Type":"ContainerDied","Data":"388afd28e6592119d8d3fd80b3b0c7fa0d1f9d0e0bb0bf2b8948007374dfc959"} Feb 27 17:32:05 crc kubenswrapper[4814]: I0227 17:32:05.443458 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="388afd28e6592119d8d3fd80b3b0c7fa0d1f9d0e0bb0bf2b8948007374dfc959" Feb 27 17:32:05 crc kubenswrapper[4814]: I0227 17:32:05.488861 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:32:05 crc kubenswrapper[4814]: E0227 17:32:05.489562 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:32:05 crc kubenswrapper[4814]: I0227 17:32:05.544702 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536892-6ktg6" Feb 27 17:32:05 crc kubenswrapper[4814]: I0227 17:32:05.596249 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r4vd\" (UniqueName: \"kubernetes.io/projected/2e9b3520-7570-4c37-8fec-58b0de0af5a9-kube-api-access-5r4vd\") pod \"2e9b3520-7570-4c37-8fec-58b0de0af5a9\" (UID: \"2e9b3520-7570-4c37-8fec-58b0de0af5a9\") " Feb 27 17:32:05 crc kubenswrapper[4814]: I0227 17:32:05.605163 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e9b3520-7570-4c37-8fec-58b0de0af5a9-kube-api-access-5r4vd" (OuterVolumeSpecName: "kube-api-access-5r4vd") pod "2e9b3520-7570-4c37-8fec-58b0de0af5a9" (UID: "2e9b3520-7570-4c37-8fec-58b0de0af5a9"). InnerVolumeSpecName "kube-api-access-5r4vd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:32:05 crc kubenswrapper[4814]: I0227 17:32:05.700193 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r4vd\" (UniqueName: \"kubernetes.io/projected/2e9b3520-7570-4c37-8fec-58b0de0af5a9-kube-api-access-5r4vd\") on node \"crc\" DevicePath \"\"" Feb 27 17:32:05 crc kubenswrapper[4814]: I0227 17:32:05.818155 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:32:05 crc kubenswrapper[4814]: I0227 17:32:05.885952 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:32:06 crc kubenswrapper[4814]: I0227 17:32:06.070224 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gpnkz"] Feb 27 17:32:06 crc kubenswrapper[4814]: I0227 17:32:06.453625 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536892-6ktg6" Feb 27 17:32:06 crc kubenswrapper[4814]: I0227 17:32:06.646643 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536886-pkn5g"] Feb 27 17:32:06 crc kubenswrapper[4814]: I0227 17:32:06.662180 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536886-pkn5g"] Feb 27 17:32:07 crc kubenswrapper[4814]: I0227 17:32:07.465957 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gpnkz" podUID="c016a833-d123-4840-ac77-72be85b6cab5" containerName="registry-server" containerID="cri-o://002186015d5fc630f91891c18f75d0e9c07e36bf888e309932081368fb45a05c" gracePeriod=2 Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.019636 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.123211 4814 scope.go:117] "RemoveContainer" containerID="fb641f670238671f623d98256379dea8bf6b61996aa684ce4b0285d944e322c0" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.169327 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c016a833-d123-4840-ac77-72be85b6cab5-catalog-content\") pod \"c016a833-d123-4840-ac77-72be85b6cab5\" (UID: \"c016a833-d123-4840-ac77-72be85b6cab5\") " Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.169542 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxnxz\" (UniqueName: \"kubernetes.io/projected/c016a833-d123-4840-ac77-72be85b6cab5-kube-api-access-lxnxz\") pod \"c016a833-d123-4840-ac77-72be85b6cab5\" (UID: \"c016a833-d123-4840-ac77-72be85b6cab5\") " Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.169705 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c016a833-d123-4840-ac77-72be85b6cab5-utilities\") pod \"c016a833-d123-4840-ac77-72be85b6cab5\" (UID: \"c016a833-d123-4840-ac77-72be85b6cab5\") " Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.172429 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c016a833-d123-4840-ac77-72be85b6cab5-utilities" (OuterVolumeSpecName: "utilities") pod "c016a833-d123-4840-ac77-72be85b6cab5" (UID: "c016a833-d123-4840-ac77-72be85b6cab5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.180685 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c016a833-d123-4840-ac77-72be85b6cab5-kube-api-access-lxnxz" (OuterVolumeSpecName: "kube-api-access-lxnxz") pod "c016a833-d123-4840-ac77-72be85b6cab5" (UID: "c016a833-d123-4840-ac77-72be85b6cab5"). InnerVolumeSpecName "kube-api-access-lxnxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.274135 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c016a833-d123-4840-ac77-72be85b6cab5-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.274210 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxnxz\" (UniqueName: \"kubernetes.io/projected/c016a833-d123-4840-ac77-72be85b6cab5-kube-api-access-lxnxz\") on node \"crc\" DevicePath \"\"" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.345227 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c016a833-d123-4840-ac77-72be85b6cab5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c016a833-d123-4840-ac77-72be85b6cab5" (UID: "c016a833-d123-4840-ac77-72be85b6cab5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.376905 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c016a833-d123-4840-ac77-72be85b6cab5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.482825 4814 generic.go:334] "Generic (PLEG): container finished" podID="c016a833-d123-4840-ac77-72be85b6cab5" containerID="002186015d5fc630f91891c18f75d0e9c07e36bf888e309932081368fb45a05c" exitCode=0 Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.482888 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gpnkz" event={"ID":"c016a833-d123-4840-ac77-72be85b6cab5","Type":"ContainerDied","Data":"002186015d5fc630f91891c18f75d0e9c07e36bf888e309932081368fb45a05c"} Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.482930 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gpnkz" event={"ID":"c016a833-d123-4840-ac77-72be85b6cab5","Type":"ContainerDied","Data":"2a4e80e975a710f4230c976a069ec46d537ba07d1d5756fb97c99874fd5fe8bf"} Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.482952 4814 scope.go:117] "RemoveContainer" containerID="002186015d5fc630f91891c18f75d0e9c07e36bf888e309932081368fb45a05c" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.482961 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gpnkz" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.508084 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7" path="/var/lib/kubelet/pods/ee243f7d-a8c8-4c2d-8dce-b5afc9ffb9b7/volumes" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.516318 4814 scope.go:117] "RemoveContainer" containerID="bc7f04db634bfd6c19ccbb2716748bb1b3c53b18fe0735f1c209d002a175aacd" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.547994 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gpnkz"] Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.563563 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gpnkz"] Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.569929 4814 scope.go:117] "RemoveContainer" containerID="2b728196e7a85b4ca25577045b89a5e205e7c858fd2982ac5a1138da36534263" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.593980 4814 scope.go:117] "RemoveContainer" containerID="002186015d5fc630f91891c18f75d0e9c07e36bf888e309932081368fb45a05c" Feb 27 17:32:08 crc kubenswrapper[4814]: E0227 17:32:08.594747 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"002186015d5fc630f91891c18f75d0e9c07e36bf888e309932081368fb45a05c\": container with ID starting with 002186015d5fc630f91891c18f75d0e9c07e36bf888e309932081368fb45a05c not found: ID does not exist" containerID="002186015d5fc630f91891c18f75d0e9c07e36bf888e309932081368fb45a05c" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.594797 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"002186015d5fc630f91891c18f75d0e9c07e36bf888e309932081368fb45a05c"} err="failed to get container status \"002186015d5fc630f91891c18f75d0e9c07e36bf888e309932081368fb45a05c\": rpc error: code = NotFound desc = could not find container \"002186015d5fc630f91891c18f75d0e9c07e36bf888e309932081368fb45a05c\": container with ID starting with 002186015d5fc630f91891c18f75d0e9c07e36bf888e309932081368fb45a05c not found: ID does not exist" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.594828 4814 scope.go:117] "RemoveContainer" containerID="bc7f04db634bfd6c19ccbb2716748bb1b3c53b18fe0735f1c209d002a175aacd" Feb 27 17:32:08 crc kubenswrapper[4814]: E0227 17:32:08.595243 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc7f04db634bfd6c19ccbb2716748bb1b3c53b18fe0735f1c209d002a175aacd\": container with ID starting with bc7f04db634bfd6c19ccbb2716748bb1b3c53b18fe0735f1c209d002a175aacd not found: ID does not exist" containerID="bc7f04db634bfd6c19ccbb2716748bb1b3c53b18fe0735f1c209d002a175aacd" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.595297 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc7f04db634bfd6c19ccbb2716748bb1b3c53b18fe0735f1c209d002a175aacd"} err="failed to get container status \"bc7f04db634bfd6c19ccbb2716748bb1b3c53b18fe0735f1c209d002a175aacd\": rpc error: code = NotFound desc = could not find container \"bc7f04db634bfd6c19ccbb2716748bb1b3c53b18fe0735f1c209d002a175aacd\": container with ID starting with bc7f04db634bfd6c19ccbb2716748bb1b3c53b18fe0735f1c209d002a175aacd not found: ID does not exist" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.595320 4814 scope.go:117] "RemoveContainer" containerID="2b728196e7a85b4ca25577045b89a5e205e7c858fd2982ac5a1138da36534263" Feb 27 17:32:08 crc kubenswrapper[4814]: E0227 17:32:08.595860 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b728196e7a85b4ca25577045b89a5e205e7c858fd2982ac5a1138da36534263\": container with ID starting with 2b728196e7a85b4ca25577045b89a5e205e7c858fd2982ac5a1138da36534263 not found: ID does not exist" containerID="2b728196e7a85b4ca25577045b89a5e205e7c858fd2982ac5a1138da36534263" Feb 27 17:32:08 crc kubenswrapper[4814]: I0227 17:32:08.595907 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b728196e7a85b4ca25577045b89a5e205e7c858fd2982ac5a1138da36534263"} err="failed to get container status \"2b728196e7a85b4ca25577045b89a5e205e7c858fd2982ac5a1138da36534263\": rpc error: code = NotFound desc = could not find container \"2b728196e7a85b4ca25577045b89a5e205e7c858fd2982ac5a1138da36534263\": container with ID starting with 2b728196e7a85b4ca25577045b89a5e205e7c858fd2982ac5a1138da36534263 not found: ID does not exist" Feb 27 17:32:10 crc kubenswrapper[4814]: I0227 17:32:10.507337 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c016a833-d123-4840-ac77-72be85b6cab5" path="/var/lib/kubelet/pods/c016a833-d123-4840-ac77-72be85b6cab5/volumes" Feb 27 17:32:14 crc kubenswrapper[4814]: E0227 17:32:14.967530 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc016a833_d123_4840_ac77_72be85b6cab5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc016a833_d123_4840_ac77_72be85b6cab5.slice/crio-2a4e80e975a710f4230c976a069ec46d537ba07d1d5756fb97c99874fd5fe8bf\": RecentStats: unable to find data in memory cache]" Feb 27 17:32:16 crc kubenswrapper[4814]: I0227 17:32:16.489009 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:32:16 crc kubenswrapper[4814]: E0227 17:32:16.489973 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:32:25 crc kubenswrapper[4814]: E0227 17:32:25.255364 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc016a833_d123_4840_ac77_72be85b6cab5.slice/crio-2a4e80e975a710f4230c976a069ec46d537ba07d1d5756fb97c99874fd5fe8bf\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc016a833_d123_4840_ac77_72be85b6cab5.slice\": RecentStats: unable to find data in memory cache]" Feb 27 17:32:31 crc kubenswrapper[4814]: I0227 17:32:31.488110 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:32:31 crc kubenswrapper[4814]: E0227 17:32:31.489655 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:32:35 crc kubenswrapper[4814]: E0227 17:32:35.553857 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc016a833_d123_4840_ac77_72be85b6cab5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc016a833_d123_4840_ac77_72be85b6cab5.slice/crio-2a4e80e975a710f4230c976a069ec46d537ba07d1d5756fb97c99874fd5fe8bf\": RecentStats: unable to find data in memory cache]" Feb 27 17:32:42 crc kubenswrapper[4814]: I0227 17:32:42.488178 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:32:42 crc kubenswrapper[4814]: E0227 17:32:42.489586 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:32:45 crc kubenswrapper[4814]: E0227 17:32:45.921562 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc016a833_d123_4840_ac77_72be85b6cab5.slice/crio-2a4e80e975a710f4230c976a069ec46d537ba07d1d5756fb97c99874fd5fe8bf\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc016a833_d123_4840_ac77_72be85b6cab5.slice\": RecentStats: unable to find data in memory cache]" Feb 27 17:32:54 crc kubenswrapper[4814]: I0227 17:32:54.488494 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:32:54 crc kubenswrapper[4814]: E0227 17:32:54.489929 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:32:56 crc kubenswrapper[4814]: E0227 17:32:56.216385 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc016a833_d123_4840_ac77_72be85b6cab5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc016a833_d123_4840_ac77_72be85b6cab5.slice/crio-2a4e80e975a710f4230c976a069ec46d537ba07d1d5756fb97c99874fd5fe8bf\": RecentStats: unable to find data in memory cache]" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.018962 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jtkl7/must-gather-28ld5"] Feb 27 17:33:05 crc kubenswrapper[4814]: E0227 17:33:05.020069 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c016a833-d123-4840-ac77-72be85b6cab5" containerName="extract-utilities" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.020085 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="c016a833-d123-4840-ac77-72be85b6cab5" containerName="extract-utilities" Feb 27 17:33:05 crc kubenswrapper[4814]: E0227 17:33:05.020099 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c016a833-d123-4840-ac77-72be85b6cab5" containerName="extract-content" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.020105 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="c016a833-d123-4840-ac77-72be85b6cab5" containerName="extract-content" Feb 27 17:33:05 crc kubenswrapper[4814]: E0227 17:33:05.020115 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e9b3520-7570-4c37-8fec-58b0de0af5a9" containerName="oc" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.020121 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e9b3520-7570-4c37-8fec-58b0de0af5a9" containerName="oc" Feb 27 17:33:05 crc kubenswrapper[4814]: E0227 17:33:05.020133 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c016a833-d123-4840-ac77-72be85b6cab5" containerName="registry-server" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.020142 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="c016a833-d123-4840-ac77-72be85b6cab5" containerName="registry-server" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.020351 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="c016a833-d123-4840-ac77-72be85b6cab5" containerName="registry-server" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.020364 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e9b3520-7570-4c37-8fec-58b0de0af5a9" containerName="oc" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.021444 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jtkl7/must-gather-28ld5" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.026140 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-jtkl7"/"default-dockercfg-mzdc4" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.026330 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jtkl7"/"openshift-service-ca.crt" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.026440 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jtkl7"/"kube-root-ca.crt" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.052868 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jtkl7/must-gather-28ld5"] Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.125587 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/22630521-470e-44ef-93dd-821e66f8bdfc-must-gather-output\") pod \"must-gather-28ld5\" (UID: \"22630521-470e-44ef-93dd-821e66f8bdfc\") " pod="openshift-must-gather-jtkl7/must-gather-28ld5" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.125735 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfcxd\" (UniqueName: \"kubernetes.io/projected/22630521-470e-44ef-93dd-821e66f8bdfc-kube-api-access-tfcxd\") pod \"must-gather-28ld5\" (UID: \"22630521-470e-44ef-93dd-821e66f8bdfc\") " pod="openshift-must-gather-jtkl7/must-gather-28ld5" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.228708 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfcxd\" (UniqueName: \"kubernetes.io/projected/22630521-470e-44ef-93dd-821e66f8bdfc-kube-api-access-tfcxd\") pod \"must-gather-28ld5\" (UID: \"22630521-470e-44ef-93dd-821e66f8bdfc\") " pod="openshift-must-gather-jtkl7/must-gather-28ld5" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.228874 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/22630521-470e-44ef-93dd-821e66f8bdfc-must-gather-output\") pod \"must-gather-28ld5\" (UID: \"22630521-470e-44ef-93dd-821e66f8bdfc\") " pod="openshift-must-gather-jtkl7/must-gather-28ld5" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.229867 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/22630521-470e-44ef-93dd-821e66f8bdfc-must-gather-output\") pod \"must-gather-28ld5\" (UID: \"22630521-470e-44ef-93dd-821e66f8bdfc\") " pod="openshift-must-gather-jtkl7/must-gather-28ld5" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.249760 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfcxd\" (UniqueName: \"kubernetes.io/projected/22630521-470e-44ef-93dd-821e66f8bdfc-kube-api-access-tfcxd\") pod \"must-gather-28ld5\" (UID: \"22630521-470e-44ef-93dd-821e66f8bdfc\") " pod="openshift-must-gather-jtkl7/must-gather-28ld5" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.352303 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jtkl7/must-gather-28ld5" Feb 27 17:33:05 crc kubenswrapper[4814]: I0227 17:33:05.862784 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jtkl7/must-gather-28ld5"] Feb 27 17:33:06 crc kubenswrapper[4814]: I0227 17:33:06.232400 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jtkl7/must-gather-28ld5" event={"ID":"22630521-470e-44ef-93dd-821e66f8bdfc","Type":"ContainerStarted","Data":"50962e7da8e64b136763b9ceaf45b3e5640ebc722b3f965efdea1113cbd82594"} Feb 27 17:33:06 crc kubenswrapper[4814]: I0227 17:33:06.232497 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jtkl7/must-gather-28ld5" event={"ID":"22630521-470e-44ef-93dd-821e66f8bdfc","Type":"ContainerStarted","Data":"8441667927b2a7712036e2a63a3a62a4b383adc96ba68403087988f3110e954a"} Feb 27 17:33:06 crc kubenswrapper[4814]: I0227 17:33:06.488386 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:33:06 crc kubenswrapper[4814]: E0227 17:33:06.489133 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:33:06 crc kubenswrapper[4814]: E0227 17:33:06.505409 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc016a833_d123_4840_ac77_72be85b6cab5.slice/crio-2a4e80e975a710f4230c976a069ec46d537ba07d1d5756fb97c99874fd5fe8bf\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc016a833_d123_4840_ac77_72be85b6cab5.slice\": RecentStats: unable to find data in memory cache]" Feb 27 17:33:07 crc kubenswrapper[4814]: I0227 17:33:07.248191 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jtkl7/must-gather-28ld5" event={"ID":"22630521-470e-44ef-93dd-821e66f8bdfc","Type":"ContainerStarted","Data":"243a98d073c905306137a55cf4a27a92a89d8e7184e8ddbc48ee54ada521d0fe"} Feb 27 17:33:07 crc kubenswrapper[4814]: I0227 17:33:07.282096 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jtkl7/must-gather-28ld5" podStartSLOduration=3.2820500360000002 podStartE2EDuration="3.282050036s" podCreationTimestamp="2026-02-27 17:33:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:33:07.273341676 +0000 UTC m=+4199.725966556" watchObservedRunningTime="2026-02-27 17:33:07.282050036 +0000 UTC m=+4199.734674906" Feb 27 17:33:08 crc kubenswrapper[4814]: I0227 17:33:08.200241 4814 scope.go:117] "RemoveContainer" containerID="f29f578709f1a3eb6fd8a6645dbd2766ce20ee9188e1615f824757758d72076d" Feb 27 17:33:10 crc kubenswrapper[4814]: I0227 17:33:10.389855 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jtkl7/crc-debug-27blh"] Feb 27 17:33:10 crc kubenswrapper[4814]: I0227 17:33:10.392501 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jtkl7/crc-debug-27blh" Feb 27 17:33:10 crc kubenswrapper[4814]: I0227 17:33:10.550606 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndz5f\" (UniqueName: \"kubernetes.io/projected/d1185daf-664b-4e8b-b1dc-91eae81d3acf-kube-api-access-ndz5f\") pod \"crc-debug-27blh\" (UID: \"d1185daf-664b-4e8b-b1dc-91eae81d3acf\") " pod="openshift-must-gather-jtkl7/crc-debug-27blh" Feb 27 17:33:10 crc kubenswrapper[4814]: I0227 17:33:10.551199 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1185daf-664b-4e8b-b1dc-91eae81d3acf-host\") pod \"crc-debug-27blh\" (UID: \"d1185daf-664b-4e8b-b1dc-91eae81d3acf\") " pod="openshift-must-gather-jtkl7/crc-debug-27blh" Feb 27 17:33:10 crc kubenswrapper[4814]: I0227 17:33:10.654252 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndz5f\" (UniqueName: \"kubernetes.io/projected/d1185daf-664b-4e8b-b1dc-91eae81d3acf-kube-api-access-ndz5f\") pod \"crc-debug-27blh\" (UID: \"d1185daf-664b-4e8b-b1dc-91eae81d3acf\") " pod="openshift-must-gather-jtkl7/crc-debug-27blh" Feb 27 17:33:10 crc kubenswrapper[4814]: I0227 17:33:10.654368 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1185daf-664b-4e8b-b1dc-91eae81d3acf-host\") pod \"crc-debug-27blh\" (UID: \"d1185daf-664b-4e8b-b1dc-91eae81d3acf\") " pod="openshift-must-gather-jtkl7/crc-debug-27blh" Feb 27 17:33:10 crc kubenswrapper[4814]: I0227 17:33:10.654627 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1185daf-664b-4e8b-b1dc-91eae81d3acf-host\") pod \"crc-debug-27blh\" (UID: \"d1185daf-664b-4e8b-b1dc-91eae81d3acf\") " pod="openshift-must-gather-jtkl7/crc-debug-27blh" Feb 27 17:33:11 crc kubenswrapper[4814]: I0227 17:33:11.477839 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndz5f\" (UniqueName: \"kubernetes.io/projected/d1185daf-664b-4e8b-b1dc-91eae81d3acf-kube-api-access-ndz5f\") pod \"crc-debug-27blh\" (UID: \"d1185daf-664b-4e8b-b1dc-91eae81d3acf\") " pod="openshift-must-gather-jtkl7/crc-debug-27blh" Feb 27 17:33:11 crc kubenswrapper[4814]: I0227 17:33:11.611499 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jtkl7/crc-debug-27blh" Feb 27 17:33:12 crc kubenswrapper[4814]: I0227 17:33:12.351521 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jtkl7/crc-debug-27blh" event={"ID":"d1185daf-664b-4e8b-b1dc-91eae81d3acf","Type":"ContainerStarted","Data":"aea93ae8fd899af60ea1a3b337c94a9243d310e2288e5f32e480b4b1239f13b7"} Feb 27 17:33:12 crc kubenswrapper[4814]: I0227 17:33:12.352291 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jtkl7/crc-debug-27blh" event={"ID":"d1185daf-664b-4e8b-b1dc-91eae81d3acf","Type":"ContainerStarted","Data":"83939db6d6055f1b3be74261458fee054c50a23f4c44eaf3117c1bf9f18a7698"} Feb 27 17:33:12 crc kubenswrapper[4814]: I0227 17:33:12.383286 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jtkl7/crc-debug-27blh" podStartSLOduration=2.383242516 podStartE2EDuration="2.383242516s" podCreationTimestamp="2026-02-27 17:33:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:33:12.371963316 +0000 UTC m=+4204.824588156" watchObservedRunningTime="2026-02-27 17:33:12.383242516 +0000 UTC m=+4204.835867366" Feb 27 17:33:20 crc kubenswrapper[4814]: I0227 17:33:20.488790 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:33:20 crc kubenswrapper[4814]: E0227 17:33:20.490110 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:33:32 crc kubenswrapper[4814]: I0227 17:33:32.488074 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:33:32 crc kubenswrapper[4814]: E0227 17:33:32.489713 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:33:47 crc kubenswrapper[4814]: I0227 17:33:47.487596 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:33:47 crc kubenswrapper[4814]: E0227 17:33:47.488540 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:33:55 crc kubenswrapper[4814]: I0227 17:33:55.023516 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cztgh"] Feb 27 17:33:55 crc kubenswrapper[4814]: I0227 17:33:55.030916 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:33:55 crc kubenswrapper[4814]: I0227 17:33:55.037005 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cztgh"] Feb 27 17:33:55 crc kubenswrapper[4814]: I0227 17:33:55.093223 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54fjw\" (UniqueName: \"kubernetes.io/projected/49940cee-0645-4a45-a7b2-05971f53dff4-kube-api-access-54fjw\") pod \"certified-operators-cztgh\" (UID: \"49940cee-0645-4a45-a7b2-05971f53dff4\") " pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:33:55 crc kubenswrapper[4814]: I0227 17:33:55.093348 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49940cee-0645-4a45-a7b2-05971f53dff4-utilities\") pod \"certified-operators-cztgh\" (UID: \"49940cee-0645-4a45-a7b2-05971f53dff4\") " pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:33:55 crc kubenswrapper[4814]: I0227 17:33:55.093417 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49940cee-0645-4a45-a7b2-05971f53dff4-catalog-content\") pod \"certified-operators-cztgh\" (UID: \"49940cee-0645-4a45-a7b2-05971f53dff4\") " pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:33:55 crc kubenswrapper[4814]: I0227 17:33:55.195831 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54fjw\" (UniqueName: \"kubernetes.io/projected/49940cee-0645-4a45-a7b2-05971f53dff4-kube-api-access-54fjw\") pod \"certified-operators-cztgh\" (UID: \"49940cee-0645-4a45-a7b2-05971f53dff4\") " pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:33:55 crc kubenswrapper[4814]: I0227 17:33:55.195893 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49940cee-0645-4a45-a7b2-05971f53dff4-utilities\") pod \"certified-operators-cztgh\" (UID: \"49940cee-0645-4a45-a7b2-05971f53dff4\") " pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:33:55 crc kubenswrapper[4814]: I0227 17:33:55.195953 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49940cee-0645-4a45-a7b2-05971f53dff4-catalog-content\") pod \"certified-operators-cztgh\" (UID: \"49940cee-0645-4a45-a7b2-05971f53dff4\") " pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:33:55 crc kubenswrapper[4814]: I0227 17:33:55.196545 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49940cee-0645-4a45-a7b2-05971f53dff4-catalog-content\") pod \"certified-operators-cztgh\" (UID: \"49940cee-0645-4a45-a7b2-05971f53dff4\") " pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:33:55 crc kubenswrapper[4814]: I0227 17:33:55.196646 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49940cee-0645-4a45-a7b2-05971f53dff4-utilities\") pod \"certified-operators-cztgh\" (UID: \"49940cee-0645-4a45-a7b2-05971f53dff4\") " pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:33:55 crc kubenswrapper[4814]: I0227 17:33:55.218711 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54fjw\" (UniqueName: \"kubernetes.io/projected/49940cee-0645-4a45-a7b2-05971f53dff4-kube-api-access-54fjw\") pod \"certified-operators-cztgh\" (UID: \"49940cee-0645-4a45-a7b2-05971f53dff4\") " pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:33:55 crc kubenswrapper[4814]: I0227 17:33:55.371686 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:33:55 crc kubenswrapper[4814]: I0227 17:33:55.945120 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cztgh"] Feb 27 17:33:56 crc kubenswrapper[4814]: I0227 17:33:56.805295 4814 generic.go:334] "Generic (PLEG): container finished" podID="49940cee-0645-4a45-a7b2-05971f53dff4" containerID="a93fdda468ca37ba99f6f0cc42ff62bb3339ac7b5306b78e1b7c916133966caf" exitCode=0 Feb 27 17:33:56 crc kubenswrapper[4814]: I0227 17:33:56.805348 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cztgh" event={"ID":"49940cee-0645-4a45-a7b2-05971f53dff4","Type":"ContainerDied","Data":"a93fdda468ca37ba99f6f0cc42ff62bb3339ac7b5306b78e1b7c916133966caf"} Feb 27 17:33:56 crc kubenswrapper[4814]: I0227 17:33:56.805694 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cztgh" event={"ID":"49940cee-0645-4a45-a7b2-05971f53dff4","Type":"ContainerStarted","Data":"a667e732b03dcd2332ebd477c6cdd0929ec485c72d4a7b1b095014ae92da3d98"} Feb 27 17:33:58 crc kubenswrapper[4814]: I0227 17:33:58.828847 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cztgh" event={"ID":"49940cee-0645-4a45-a7b2-05971f53dff4","Type":"ContainerStarted","Data":"b9edc454bb284b105b1502725effde56bc132b24078e756d4eebac44df4daaf5"} Feb 27 17:33:59 crc kubenswrapper[4814]: I0227 17:33:59.488097 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:33:59 crc kubenswrapper[4814]: I0227 17:33:59.846308 4814 generic.go:334] "Generic (PLEG): container finished" podID="49940cee-0645-4a45-a7b2-05971f53dff4" containerID="b9edc454bb284b105b1502725effde56bc132b24078e756d4eebac44df4daaf5" exitCode=0 Feb 27 17:33:59 crc kubenswrapper[4814]: I0227 17:33:59.846561 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cztgh" event={"ID":"49940cee-0645-4a45-a7b2-05971f53dff4","Type":"ContainerDied","Data":"b9edc454bb284b105b1502725effde56bc132b24078e756d4eebac44df4daaf5"} Feb 27 17:33:59 crc kubenswrapper[4814]: I0227 17:33:59.849519 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"1c245446ce2fe408fb6882fd53824f92e0a19fb4974ea6eaf11f89bd2096f9b4"} Feb 27 17:34:00 crc kubenswrapper[4814]: I0227 17:34:00.150787 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536894-f45jh"] Feb 27 17:34:00 crc kubenswrapper[4814]: I0227 17:34:00.152116 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536894-f45jh" Feb 27 17:34:00 crc kubenswrapper[4814]: I0227 17:34:00.155837 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:34:00 crc kubenswrapper[4814]: I0227 17:34:00.156136 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:34:00 crc kubenswrapper[4814]: I0227 17:34:00.156360 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:34:00 crc kubenswrapper[4814]: I0227 17:34:00.162290 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536894-f45jh"] Feb 27 17:34:00 crc kubenswrapper[4814]: I0227 17:34:00.230246 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm74s\" (UniqueName: \"kubernetes.io/projected/5964bc58-dabe-4e8f-859e-824d177aa798-kube-api-access-rm74s\") pod \"auto-csr-approver-29536894-f45jh\" (UID: \"5964bc58-dabe-4e8f-859e-824d177aa798\") " pod="openshift-infra/auto-csr-approver-29536894-f45jh" Feb 27 17:34:00 crc kubenswrapper[4814]: I0227 17:34:00.332754 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm74s\" (UniqueName: \"kubernetes.io/projected/5964bc58-dabe-4e8f-859e-824d177aa798-kube-api-access-rm74s\") pod \"auto-csr-approver-29536894-f45jh\" (UID: \"5964bc58-dabe-4e8f-859e-824d177aa798\") " pod="openshift-infra/auto-csr-approver-29536894-f45jh" Feb 27 17:34:00 crc kubenswrapper[4814]: I0227 17:34:00.360113 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm74s\" (UniqueName: \"kubernetes.io/projected/5964bc58-dabe-4e8f-859e-824d177aa798-kube-api-access-rm74s\") pod \"auto-csr-approver-29536894-f45jh\" (UID: \"5964bc58-dabe-4e8f-859e-824d177aa798\") " pod="openshift-infra/auto-csr-approver-29536894-f45jh" Feb 27 17:34:00 crc kubenswrapper[4814]: I0227 17:34:00.474291 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536894-f45jh" Feb 27 17:34:00 crc kubenswrapper[4814]: I0227 17:34:00.862089 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cztgh" event={"ID":"49940cee-0645-4a45-a7b2-05971f53dff4","Type":"ContainerStarted","Data":"c98b5e848d5821a92eb38c61aa505b064e827f2f054d81e3771544199716e8b4"} Feb 27 17:34:00 crc kubenswrapper[4814]: I0227 17:34:00.891640 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cztgh" podStartSLOduration=2.42432094 podStartE2EDuration="5.891615714s" podCreationTimestamp="2026-02-27 17:33:55 +0000 UTC" firstStartedPulling="2026-02-27 17:33:56.808440453 +0000 UTC m=+4249.261065283" lastFinishedPulling="2026-02-27 17:34:00.275735227 +0000 UTC m=+4252.728360057" observedRunningTime="2026-02-27 17:34:00.883174672 +0000 UTC m=+4253.335799492" watchObservedRunningTime="2026-02-27 17:34:00.891615714 +0000 UTC m=+4253.344240544" Feb 27 17:34:00 crc kubenswrapper[4814]: I0227 17:34:00.959195 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536894-f45jh"] Feb 27 17:34:01 crc kubenswrapper[4814]: I0227 17:34:01.874822 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536894-f45jh" event={"ID":"5964bc58-dabe-4e8f-859e-824d177aa798","Type":"ContainerStarted","Data":"e17da29205c8e3d7eef75a87b1ba1d0503d12f7ff44666781d27449499da89c9"} Feb 27 17:34:02 crc kubenswrapper[4814]: I0227 17:34:02.887615 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536894-f45jh" event={"ID":"5964bc58-dabe-4e8f-859e-824d177aa798","Type":"ContainerStarted","Data":"8107e26799afa00a83f85963b78b1a0cc4fabbd17f38d15594f812c7dc61f66d"} Feb 27 17:34:02 crc kubenswrapper[4814]: I0227 17:34:02.908699 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536894-f45jh" podStartSLOduration=1.975729484 podStartE2EDuration="2.908672132s" podCreationTimestamp="2026-02-27 17:34:00 +0000 UTC" firstStartedPulling="2026-02-27 17:34:01.391529131 +0000 UTC m=+4253.844153961" lastFinishedPulling="2026-02-27 17:34:02.324471779 +0000 UTC m=+4254.777096609" observedRunningTime="2026-02-27 17:34:02.905937127 +0000 UTC m=+4255.358561947" watchObservedRunningTime="2026-02-27 17:34:02.908672132 +0000 UTC m=+4255.361296962" Feb 27 17:34:03 crc kubenswrapper[4814]: I0227 17:34:03.912326 4814 generic.go:334] "Generic (PLEG): container finished" podID="5964bc58-dabe-4e8f-859e-824d177aa798" containerID="8107e26799afa00a83f85963b78b1a0cc4fabbd17f38d15594f812c7dc61f66d" exitCode=0 Feb 27 17:34:03 crc kubenswrapper[4814]: I0227 17:34:03.912844 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536894-f45jh" event={"ID":"5964bc58-dabe-4e8f-859e-824d177aa798","Type":"ContainerDied","Data":"8107e26799afa00a83f85963b78b1a0cc4fabbd17f38d15594f812c7dc61f66d"} Feb 27 17:34:04 crc kubenswrapper[4814]: I0227 17:34:04.929621 4814 generic.go:334] "Generic (PLEG): container finished" podID="d1185daf-664b-4e8b-b1dc-91eae81d3acf" containerID="aea93ae8fd899af60ea1a3b337c94a9243d310e2288e5f32e480b4b1239f13b7" exitCode=0 Feb 27 17:34:04 crc kubenswrapper[4814]: I0227 17:34:04.929723 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jtkl7/crc-debug-27blh" event={"ID":"d1185daf-664b-4e8b-b1dc-91eae81d3acf","Type":"ContainerDied","Data":"aea93ae8fd899af60ea1a3b337c94a9243d310e2288e5f32e480b4b1239f13b7"} Feb 27 17:34:05 crc kubenswrapper[4814]: I0227 17:34:05.326988 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536894-f45jh" Feb 27 17:34:05 crc kubenswrapper[4814]: I0227 17:34:05.358426 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rm74s\" (UniqueName: \"kubernetes.io/projected/5964bc58-dabe-4e8f-859e-824d177aa798-kube-api-access-rm74s\") pod \"5964bc58-dabe-4e8f-859e-824d177aa798\" (UID: \"5964bc58-dabe-4e8f-859e-824d177aa798\") " Feb 27 17:34:05 crc kubenswrapper[4814]: I0227 17:34:05.372837 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5964bc58-dabe-4e8f-859e-824d177aa798-kube-api-access-rm74s" (OuterVolumeSpecName: "kube-api-access-rm74s") pod "5964bc58-dabe-4e8f-859e-824d177aa798" (UID: "5964bc58-dabe-4e8f-859e-824d177aa798"). InnerVolumeSpecName "kube-api-access-rm74s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:34:05 crc kubenswrapper[4814]: I0227 17:34:05.373072 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:34:05 crc kubenswrapper[4814]: I0227 17:34:05.373943 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:34:05 crc kubenswrapper[4814]: I0227 17:34:05.464262 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rm74s\" (UniqueName: \"kubernetes.io/projected/5964bc58-dabe-4e8f-859e-824d177aa798-kube-api-access-rm74s\") on node \"crc\" DevicePath \"\"" Feb 27 17:34:05 crc kubenswrapper[4814]: I0227 17:34:05.505613 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:34:05 crc kubenswrapper[4814]: I0227 17:34:05.942987 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536894-f45jh" event={"ID":"5964bc58-dabe-4e8f-859e-824d177aa798","Type":"ContainerDied","Data":"e17da29205c8e3d7eef75a87b1ba1d0503d12f7ff44666781d27449499da89c9"} Feb 27 17:34:05 crc kubenswrapper[4814]: I0227 17:34:05.943037 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536894-f45jh" Feb 27 17:34:05 crc kubenswrapper[4814]: I0227 17:34:05.943055 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e17da29205c8e3d7eef75a87b1ba1d0503d12f7ff44666781d27449499da89c9" Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.008843 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536888-mcrll"] Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.015396 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.022763 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536888-mcrll"] Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.057048 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jtkl7/crc-debug-27blh" Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.076156 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cztgh"] Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.102436 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jtkl7/crc-debug-27blh"] Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.113380 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jtkl7/crc-debug-27blh"] Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.178780 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndz5f\" (UniqueName: \"kubernetes.io/projected/d1185daf-664b-4e8b-b1dc-91eae81d3acf-kube-api-access-ndz5f\") pod \"d1185daf-664b-4e8b-b1dc-91eae81d3acf\" (UID: \"d1185daf-664b-4e8b-b1dc-91eae81d3acf\") " Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.179050 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1185daf-664b-4e8b-b1dc-91eae81d3acf-host\") pod \"d1185daf-664b-4e8b-b1dc-91eae81d3acf\" (UID: \"d1185daf-664b-4e8b-b1dc-91eae81d3acf\") " Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.179419 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1185daf-664b-4e8b-b1dc-91eae81d3acf-host" (OuterVolumeSpecName: "host") pod "d1185daf-664b-4e8b-b1dc-91eae81d3acf" (UID: "d1185daf-664b-4e8b-b1dc-91eae81d3acf"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.179703 4814 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1185daf-664b-4e8b-b1dc-91eae81d3acf-host\") on node \"crc\" DevicePath \"\"" Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.185760 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1185daf-664b-4e8b-b1dc-91eae81d3acf-kube-api-access-ndz5f" (OuterVolumeSpecName: "kube-api-access-ndz5f") pod "d1185daf-664b-4e8b-b1dc-91eae81d3acf" (UID: "d1185daf-664b-4e8b-b1dc-91eae81d3acf"). InnerVolumeSpecName "kube-api-access-ndz5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.281855 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndz5f\" (UniqueName: \"kubernetes.io/projected/d1185daf-664b-4e8b-b1dc-91eae81d3acf-kube-api-access-ndz5f\") on node \"crc\" DevicePath \"\"" Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.509032 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1185daf-664b-4e8b-b1dc-91eae81d3acf" path="/var/lib/kubelet/pods/d1185daf-664b-4e8b-b1dc-91eae81d3acf/volumes" Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.509828 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbea8566-2a73-43f6-9b57-8372114149fb" path="/var/lib/kubelet/pods/fbea8566-2a73-43f6-9b57-8372114149fb/volumes" Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.958649 4814 scope.go:117] "RemoveContainer" containerID="aea93ae8fd899af60ea1a3b337c94a9243d310e2288e5f32e480b4b1239f13b7" Feb 27 17:34:06 crc kubenswrapper[4814]: I0227 17:34:06.958691 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jtkl7/crc-debug-27blh" Feb 27 17:34:07 crc kubenswrapper[4814]: I0227 17:34:07.361392 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jtkl7/crc-debug-qt2gm"] Feb 27 17:34:07 crc kubenswrapper[4814]: E0227 17:34:07.362148 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5964bc58-dabe-4e8f-859e-824d177aa798" containerName="oc" Feb 27 17:34:07 crc kubenswrapper[4814]: I0227 17:34:07.362181 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="5964bc58-dabe-4e8f-859e-824d177aa798" containerName="oc" Feb 27 17:34:07 crc kubenswrapper[4814]: E0227 17:34:07.362242 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1185daf-664b-4e8b-b1dc-91eae81d3acf" containerName="container-00" Feb 27 17:34:07 crc kubenswrapper[4814]: I0227 17:34:07.362291 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1185daf-664b-4e8b-b1dc-91eae81d3acf" containerName="container-00" Feb 27 17:34:07 crc kubenswrapper[4814]: I0227 17:34:07.362739 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="5964bc58-dabe-4e8f-859e-824d177aa798" containerName="oc" Feb 27 17:34:07 crc kubenswrapper[4814]: I0227 17:34:07.362804 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1185daf-664b-4e8b-b1dc-91eae81d3acf" containerName="container-00" Feb 27 17:34:07 crc kubenswrapper[4814]: I0227 17:34:07.365589 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jtkl7/crc-debug-qt2gm" Feb 27 17:34:07 crc kubenswrapper[4814]: I0227 17:34:07.412324 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c28f0ae-fab1-49ee-b184-23c1ef2b852d-host\") pod \"crc-debug-qt2gm\" (UID: \"9c28f0ae-fab1-49ee-b184-23c1ef2b852d\") " pod="openshift-must-gather-jtkl7/crc-debug-qt2gm" Feb 27 17:34:07 crc kubenswrapper[4814]: I0227 17:34:07.412564 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22qlb\" (UniqueName: \"kubernetes.io/projected/9c28f0ae-fab1-49ee-b184-23c1ef2b852d-kube-api-access-22qlb\") pod \"crc-debug-qt2gm\" (UID: \"9c28f0ae-fab1-49ee-b184-23c1ef2b852d\") " pod="openshift-must-gather-jtkl7/crc-debug-qt2gm" Feb 27 17:34:07 crc kubenswrapper[4814]: I0227 17:34:07.515215 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c28f0ae-fab1-49ee-b184-23c1ef2b852d-host\") pod \"crc-debug-qt2gm\" (UID: \"9c28f0ae-fab1-49ee-b184-23c1ef2b852d\") " pod="openshift-must-gather-jtkl7/crc-debug-qt2gm" Feb 27 17:34:07 crc kubenswrapper[4814]: I0227 17:34:07.515452 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22qlb\" (UniqueName: \"kubernetes.io/projected/9c28f0ae-fab1-49ee-b184-23c1ef2b852d-kube-api-access-22qlb\") pod \"crc-debug-qt2gm\" (UID: \"9c28f0ae-fab1-49ee-b184-23c1ef2b852d\") " pod="openshift-must-gather-jtkl7/crc-debug-qt2gm" Feb 27 17:34:07 crc kubenswrapper[4814]: I0227 17:34:07.515798 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c28f0ae-fab1-49ee-b184-23c1ef2b852d-host\") pod \"crc-debug-qt2gm\" (UID: \"9c28f0ae-fab1-49ee-b184-23c1ef2b852d\") " pod="openshift-must-gather-jtkl7/crc-debug-qt2gm" Feb 27 17:34:07 crc kubenswrapper[4814]: I0227 17:34:07.543181 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22qlb\" (UniqueName: \"kubernetes.io/projected/9c28f0ae-fab1-49ee-b184-23c1ef2b852d-kube-api-access-22qlb\") pod \"crc-debug-qt2gm\" (UID: \"9c28f0ae-fab1-49ee-b184-23c1ef2b852d\") " pod="openshift-must-gather-jtkl7/crc-debug-qt2gm" Feb 27 17:34:07 crc kubenswrapper[4814]: I0227 17:34:07.709593 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jtkl7/crc-debug-qt2gm" Feb 27 17:34:07 crc kubenswrapper[4814]: I0227 17:34:07.971806 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jtkl7/crc-debug-qt2gm" event={"ID":"9c28f0ae-fab1-49ee-b184-23c1ef2b852d","Type":"ContainerStarted","Data":"99a2e05caff3e1114ac93eb09e6f4eb8ce6d593c8731b7dbf4ffb51114b362e2"} Feb 27 17:34:07 crc kubenswrapper[4814]: I0227 17:34:07.974979 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cztgh" podUID="49940cee-0645-4a45-a7b2-05971f53dff4" containerName="registry-server" containerID="cri-o://c98b5e848d5821a92eb38c61aa505b064e827f2f054d81e3771544199716e8b4" gracePeriod=2 Feb 27 17:34:08 crc kubenswrapper[4814]: E0227 17:34:08.112805 4814 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49940cee_0645_4a45_a7b2_05971f53dff4.slice/crio-conmon-c98b5e848d5821a92eb38c61aa505b064e827f2f054d81e3771544199716e8b4.scope\": RecentStats: unable to find data in memory cache]" Feb 27 17:34:08 crc kubenswrapper[4814]: I0227 17:34:08.367809 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:34:08 crc kubenswrapper[4814]: I0227 17:34:08.440034 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49940cee-0645-4a45-a7b2-05971f53dff4-catalog-content\") pod \"49940cee-0645-4a45-a7b2-05971f53dff4\" (UID: \"49940cee-0645-4a45-a7b2-05971f53dff4\") " Feb 27 17:34:08 crc kubenswrapper[4814]: I0227 17:34:08.440376 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54fjw\" (UniqueName: \"kubernetes.io/projected/49940cee-0645-4a45-a7b2-05971f53dff4-kube-api-access-54fjw\") pod \"49940cee-0645-4a45-a7b2-05971f53dff4\" (UID: \"49940cee-0645-4a45-a7b2-05971f53dff4\") " Feb 27 17:34:08 crc kubenswrapper[4814]: I0227 17:34:08.440499 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49940cee-0645-4a45-a7b2-05971f53dff4-utilities\") pod \"49940cee-0645-4a45-a7b2-05971f53dff4\" (UID: \"49940cee-0645-4a45-a7b2-05971f53dff4\") " Feb 27 17:34:08 crc kubenswrapper[4814]: I0227 17:34:08.441667 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49940cee-0645-4a45-a7b2-05971f53dff4-utilities" (OuterVolumeSpecName: "utilities") pod "49940cee-0645-4a45-a7b2-05971f53dff4" (UID: "49940cee-0645-4a45-a7b2-05971f53dff4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:34:08 crc kubenswrapper[4814]: I0227 17:34:08.448525 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49940cee-0645-4a45-a7b2-05971f53dff4-kube-api-access-54fjw" (OuterVolumeSpecName: "kube-api-access-54fjw") pod "49940cee-0645-4a45-a7b2-05971f53dff4" (UID: "49940cee-0645-4a45-a7b2-05971f53dff4"). InnerVolumeSpecName "kube-api-access-54fjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:34:08 crc kubenswrapper[4814]: I0227 17:34:08.546369 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54fjw\" (UniqueName: \"kubernetes.io/projected/49940cee-0645-4a45-a7b2-05971f53dff4-kube-api-access-54fjw\") on node \"crc\" DevicePath \"\"" Feb 27 17:34:08 crc kubenswrapper[4814]: I0227 17:34:08.546404 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49940cee-0645-4a45-a7b2-05971f53dff4-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:34:08 crc kubenswrapper[4814]: I0227 17:34:08.992786 4814 generic.go:334] "Generic (PLEG): container finished" podID="49940cee-0645-4a45-a7b2-05971f53dff4" containerID="c98b5e848d5821a92eb38c61aa505b064e827f2f054d81e3771544199716e8b4" exitCode=0 Feb 27 17:34:08 crc kubenswrapper[4814]: I0227 17:34:08.992955 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cztgh" Feb 27 17:34:08 crc kubenswrapper[4814]: I0227 17:34:08.992945 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cztgh" event={"ID":"49940cee-0645-4a45-a7b2-05971f53dff4","Type":"ContainerDied","Data":"c98b5e848d5821a92eb38c61aa505b064e827f2f054d81e3771544199716e8b4"} Feb 27 17:34:08 crc kubenswrapper[4814]: I0227 17:34:08.994484 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cztgh" event={"ID":"49940cee-0645-4a45-a7b2-05971f53dff4","Type":"ContainerDied","Data":"a667e732b03dcd2332ebd477c6cdd0929ec485c72d4a7b1b095014ae92da3d98"} Feb 27 17:34:08 crc kubenswrapper[4814]: I0227 17:34:08.994570 4814 scope.go:117] "RemoveContainer" containerID="c98b5e848d5821a92eb38c61aa505b064e827f2f054d81e3771544199716e8b4" Feb 27 17:34:08 crc kubenswrapper[4814]: I0227 17:34:08.997023 4814 generic.go:334] "Generic (PLEG): container finished" podID="9c28f0ae-fab1-49ee-b184-23c1ef2b852d" containerID="53ff9412dbbc82927534231ff428af77d77bbecf6d27a5db8f8a78f535389f55" exitCode=0 Feb 27 17:34:08 crc kubenswrapper[4814]: I0227 17:34:08.997101 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jtkl7/crc-debug-qt2gm" event={"ID":"9c28f0ae-fab1-49ee-b184-23c1ef2b852d","Type":"ContainerDied","Data":"53ff9412dbbc82927534231ff428af77d77bbecf6d27a5db8f8a78f535389f55"} Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.029024 4814 scope.go:117] "RemoveContainer" containerID="b9edc454bb284b105b1502725effde56bc132b24078e756d4eebac44df4daaf5" Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.068063 4814 scope.go:117] "RemoveContainer" containerID="a93fdda468ca37ba99f6f0cc42ff62bb3339ac7b5306b78e1b7c916133966caf" Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.143808 4814 scope.go:117] "RemoveContainer" containerID="e5e285ac2f7eaff66116ffcb22e228c0831edba91d524019e361e3bdfedaa28b" Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.148681 4814 scope.go:117] "RemoveContainer" containerID="c98b5e848d5821a92eb38c61aa505b064e827f2f054d81e3771544199716e8b4" Feb 27 17:34:09 crc kubenswrapper[4814]: E0227 17:34:09.151143 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c98b5e848d5821a92eb38c61aa505b064e827f2f054d81e3771544199716e8b4\": container with ID starting with c98b5e848d5821a92eb38c61aa505b064e827f2f054d81e3771544199716e8b4 not found: ID does not exist" containerID="c98b5e848d5821a92eb38c61aa505b064e827f2f054d81e3771544199716e8b4" Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.151201 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c98b5e848d5821a92eb38c61aa505b064e827f2f054d81e3771544199716e8b4"} err="failed to get container status \"c98b5e848d5821a92eb38c61aa505b064e827f2f054d81e3771544199716e8b4\": rpc error: code = NotFound desc = could not find container \"c98b5e848d5821a92eb38c61aa505b064e827f2f054d81e3771544199716e8b4\": container with ID starting with c98b5e848d5821a92eb38c61aa505b064e827f2f054d81e3771544199716e8b4 not found: ID does not exist" Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.151241 4814 scope.go:117] "RemoveContainer" containerID="b9edc454bb284b105b1502725effde56bc132b24078e756d4eebac44df4daaf5" Feb 27 17:34:09 crc kubenswrapper[4814]: E0227 17:34:09.152533 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9edc454bb284b105b1502725effde56bc132b24078e756d4eebac44df4daaf5\": container with ID starting with b9edc454bb284b105b1502725effde56bc132b24078e756d4eebac44df4daaf5 not found: ID does not exist" containerID="b9edc454bb284b105b1502725effde56bc132b24078e756d4eebac44df4daaf5" Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.152623 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9edc454bb284b105b1502725effde56bc132b24078e756d4eebac44df4daaf5"} err="failed to get container status \"b9edc454bb284b105b1502725effde56bc132b24078e756d4eebac44df4daaf5\": rpc error: code = NotFound desc = could not find container \"b9edc454bb284b105b1502725effde56bc132b24078e756d4eebac44df4daaf5\": container with ID starting with b9edc454bb284b105b1502725effde56bc132b24078e756d4eebac44df4daaf5 not found: ID does not exist" Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.152674 4814 scope.go:117] "RemoveContainer" containerID="a93fdda468ca37ba99f6f0cc42ff62bb3339ac7b5306b78e1b7c916133966caf" Feb 27 17:34:09 crc kubenswrapper[4814]: E0227 17:34:09.153666 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a93fdda468ca37ba99f6f0cc42ff62bb3339ac7b5306b78e1b7c916133966caf\": container with ID starting with a93fdda468ca37ba99f6f0cc42ff62bb3339ac7b5306b78e1b7c916133966caf not found: ID does not exist" containerID="a93fdda468ca37ba99f6f0cc42ff62bb3339ac7b5306b78e1b7c916133966caf" Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.153733 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a93fdda468ca37ba99f6f0cc42ff62bb3339ac7b5306b78e1b7c916133966caf"} err="failed to get container status \"a93fdda468ca37ba99f6f0cc42ff62bb3339ac7b5306b78e1b7c916133966caf\": rpc error: code = NotFound desc = could not find container \"a93fdda468ca37ba99f6f0cc42ff62bb3339ac7b5306b78e1b7c916133966caf\": container with ID starting with a93fdda468ca37ba99f6f0cc42ff62bb3339ac7b5306b78e1b7c916133966caf not found: ID does not exist" Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.261656 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49940cee-0645-4a45-a7b2-05971f53dff4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49940cee-0645-4a45-a7b2-05971f53dff4" (UID: "49940cee-0645-4a45-a7b2-05971f53dff4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.263158 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49940cee-0645-4a45-a7b2-05971f53dff4-catalog-content\") pod \"49940cee-0645-4a45-a7b2-05971f53dff4\" (UID: \"49940cee-0645-4a45-a7b2-05971f53dff4\") " Feb 27 17:34:09 crc kubenswrapper[4814]: W0227 17:34:09.263367 4814 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/49940cee-0645-4a45-a7b2-05971f53dff4/volumes/kubernetes.io~empty-dir/catalog-content Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.263511 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49940cee-0645-4a45-a7b2-05971f53dff4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49940cee-0645-4a45-a7b2-05971f53dff4" (UID: "49940cee-0645-4a45-a7b2-05971f53dff4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.264146 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49940cee-0645-4a45-a7b2-05971f53dff4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.336315 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cztgh"] Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.353558 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cztgh"] Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.639200 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jtkl7/crc-debug-qt2gm"] Feb 27 17:34:09 crc kubenswrapper[4814]: I0227 17:34:09.651207 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jtkl7/crc-debug-qt2gm"] Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.158042 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jtkl7/crc-debug-qt2gm" Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.297131 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c28f0ae-fab1-49ee-b184-23c1ef2b852d-host\") pod \"9c28f0ae-fab1-49ee-b184-23c1ef2b852d\" (UID: \"9c28f0ae-fab1-49ee-b184-23c1ef2b852d\") " Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.297373 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c28f0ae-fab1-49ee-b184-23c1ef2b852d-host" (OuterVolumeSpecName: "host") pod "9c28f0ae-fab1-49ee-b184-23c1ef2b852d" (UID: "9c28f0ae-fab1-49ee-b184-23c1ef2b852d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.297946 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22qlb\" (UniqueName: \"kubernetes.io/projected/9c28f0ae-fab1-49ee-b184-23c1ef2b852d-kube-api-access-22qlb\") pod \"9c28f0ae-fab1-49ee-b184-23c1ef2b852d\" (UID: \"9c28f0ae-fab1-49ee-b184-23c1ef2b852d\") " Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.299035 4814 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c28f0ae-fab1-49ee-b184-23c1ef2b852d-host\") on node \"crc\" DevicePath \"\"" Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.308232 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c28f0ae-fab1-49ee-b184-23c1ef2b852d-kube-api-access-22qlb" (OuterVolumeSpecName: "kube-api-access-22qlb") pod "9c28f0ae-fab1-49ee-b184-23c1ef2b852d" (UID: "9c28f0ae-fab1-49ee-b184-23c1ef2b852d"). InnerVolumeSpecName "kube-api-access-22qlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.402118 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22qlb\" (UniqueName: \"kubernetes.io/projected/9c28f0ae-fab1-49ee-b184-23c1ef2b852d-kube-api-access-22qlb\") on node \"crc\" DevicePath \"\"" Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.512431 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49940cee-0645-4a45-a7b2-05971f53dff4" path="/var/lib/kubelet/pods/49940cee-0645-4a45-a7b2-05971f53dff4/volumes" Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.515168 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c28f0ae-fab1-49ee-b184-23c1ef2b852d" path="/var/lib/kubelet/pods/9c28f0ae-fab1-49ee-b184-23c1ef2b852d/volumes" Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.869035 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jtkl7/crc-debug-58k7f"] Feb 27 17:34:10 crc kubenswrapper[4814]: E0227 17:34:10.870469 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49940cee-0645-4a45-a7b2-05971f53dff4" containerName="extract-utilities" Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.870498 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="49940cee-0645-4a45-a7b2-05971f53dff4" containerName="extract-utilities" Feb 27 17:34:10 crc kubenswrapper[4814]: E0227 17:34:10.870536 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49940cee-0645-4a45-a7b2-05971f53dff4" containerName="registry-server" Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.870548 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="49940cee-0645-4a45-a7b2-05971f53dff4" containerName="registry-server" Feb 27 17:34:10 crc kubenswrapper[4814]: E0227 17:34:10.870608 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49940cee-0645-4a45-a7b2-05971f53dff4" containerName="extract-content" Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.870618 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="49940cee-0645-4a45-a7b2-05971f53dff4" containerName="extract-content" Feb 27 17:34:10 crc kubenswrapper[4814]: E0227 17:34:10.870628 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c28f0ae-fab1-49ee-b184-23c1ef2b852d" containerName="container-00" Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.870638 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c28f0ae-fab1-49ee-b184-23c1ef2b852d" containerName="container-00" Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.870910 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c28f0ae-fab1-49ee-b184-23c1ef2b852d" containerName="container-00" Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.870973 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="49940cee-0645-4a45-a7b2-05971f53dff4" containerName="registry-server" Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.872070 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jtkl7/crc-debug-58k7f" Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.925979 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d-host\") pod \"crc-debug-58k7f\" (UID: \"adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d\") " pod="openshift-must-gather-jtkl7/crc-debug-58k7f" Feb 27 17:34:10 crc kubenswrapper[4814]: I0227 17:34:10.926300 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fmlr\" (UniqueName: \"kubernetes.io/projected/adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d-kube-api-access-7fmlr\") pod \"crc-debug-58k7f\" (UID: \"adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d\") " pod="openshift-must-gather-jtkl7/crc-debug-58k7f" Feb 27 17:34:11 crc kubenswrapper[4814]: I0227 17:34:11.028792 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fmlr\" (UniqueName: \"kubernetes.io/projected/adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d-kube-api-access-7fmlr\") pod \"crc-debug-58k7f\" (UID: \"adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d\") " pod="openshift-must-gather-jtkl7/crc-debug-58k7f" Feb 27 17:34:11 crc kubenswrapper[4814]: I0227 17:34:11.028913 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d-host\") pod \"crc-debug-58k7f\" (UID: \"adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d\") " pod="openshift-must-gather-jtkl7/crc-debug-58k7f" Feb 27 17:34:11 crc kubenswrapper[4814]: I0227 17:34:11.029071 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d-host\") pod \"crc-debug-58k7f\" (UID: \"adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d\") " pod="openshift-must-gather-jtkl7/crc-debug-58k7f" Feb 27 17:34:11 crc kubenswrapper[4814]: I0227 17:34:11.038242 4814 scope.go:117] "RemoveContainer" containerID="53ff9412dbbc82927534231ff428af77d77bbecf6d27a5db8f8a78f535389f55" Feb 27 17:34:11 crc kubenswrapper[4814]: I0227 17:34:11.038349 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jtkl7/crc-debug-qt2gm" Feb 27 17:34:11 crc kubenswrapper[4814]: I0227 17:34:11.052778 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fmlr\" (UniqueName: \"kubernetes.io/projected/adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d-kube-api-access-7fmlr\") pod \"crc-debug-58k7f\" (UID: \"adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d\") " pod="openshift-must-gather-jtkl7/crc-debug-58k7f" Feb 27 17:34:11 crc kubenswrapper[4814]: I0227 17:34:11.208017 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jtkl7/crc-debug-58k7f" Feb 27 17:34:11 crc kubenswrapper[4814]: W0227 17:34:11.252802 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadf2e1ff_6a6e_46e3_a8ba_4d0f575a748d.slice/crio-9edee11f227400c81572d45c89e1e99aa733e57591be95b5c28501b4654afeb5 WatchSource:0}: Error finding container 9edee11f227400c81572d45c89e1e99aa733e57591be95b5c28501b4654afeb5: Status 404 returned error can't find the container with id 9edee11f227400c81572d45c89e1e99aa733e57591be95b5c28501b4654afeb5 Feb 27 17:34:12 crc kubenswrapper[4814]: I0227 17:34:12.050430 4814 generic.go:334] "Generic (PLEG): container finished" podID="adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d" containerID="c185adda0c3dd948a8e9d76d6a7f2c076314ac2df03c015c1716cfe1f765d6ef" exitCode=0 Feb 27 17:34:12 crc kubenswrapper[4814]: I0227 17:34:12.050541 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jtkl7/crc-debug-58k7f" event={"ID":"adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d","Type":"ContainerDied","Data":"c185adda0c3dd948a8e9d76d6a7f2c076314ac2df03c015c1716cfe1f765d6ef"} Feb 27 17:34:12 crc kubenswrapper[4814]: I0227 17:34:12.050628 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jtkl7/crc-debug-58k7f" event={"ID":"adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d","Type":"ContainerStarted","Data":"9edee11f227400c81572d45c89e1e99aa733e57591be95b5c28501b4654afeb5"} Feb 27 17:34:12 crc kubenswrapper[4814]: I0227 17:34:12.102598 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jtkl7/crc-debug-58k7f"] Feb 27 17:34:12 crc kubenswrapper[4814]: I0227 17:34:12.114412 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jtkl7/crc-debug-58k7f"] Feb 27 17:34:13 crc kubenswrapper[4814]: I0227 17:34:13.158020 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jtkl7/crc-debug-58k7f" Feb 27 17:34:13 crc kubenswrapper[4814]: I0227 17:34:13.287162 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d-host\") pod \"adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d\" (UID: \"adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d\") " Feb 27 17:34:13 crc kubenswrapper[4814]: I0227 17:34:13.287375 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d-host" (OuterVolumeSpecName: "host") pod "adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d" (UID: "adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:34:13 crc kubenswrapper[4814]: I0227 17:34:13.287450 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fmlr\" (UniqueName: \"kubernetes.io/projected/adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d-kube-api-access-7fmlr\") pod \"adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d\" (UID: \"adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d\") " Feb 27 17:34:13 crc kubenswrapper[4814]: I0227 17:34:13.288446 4814 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d-host\") on node \"crc\" DevicePath \"\"" Feb 27 17:34:13 crc kubenswrapper[4814]: I0227 17:34:13.314035 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d-kube-api-access-7fmlr" (OuterVolumeSpecName: "kube-api-access-7fmlr") pod "adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d" (UID: "adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d"). InnerVolumeSpecName "kube-api-access-7fmlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:34:13 crc kubenswrapper[4814]: I0227 17:34:13.391503 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fmlr\" (UniqueName: \"kubernetes.io/projected/adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d-kube-api-access-7fmlr\") on node \"crc\" DevicePath \"\"" Feb 27 17:34:14 crc kubenswrapper[4814]: I0227 17:34:14.074801 4814 scope.go:117] "RemoveContainer" containerID="c185adda0c3dd948a8e9d76d6a7f2c076314ac2df03c015c1716cfe1f765d6ef" Feb 27 17:34:14 crc kubenswrapper[4814]: I0227 17:34:14.074965 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jtkl7/crc-debug-58k7f" Feb 27 17:34:14 crc kubenswrapper[4814]: I0227 17:34:14.518968 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d" path="/var/lib/kubelet/pods/adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d/volumes" Feb 27 17:34:36 crc kubenswrapper[4814]: I0227 17:34:36.616029 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-f8687c5bd-zdnxs_1121d7e8-a956-4a6b-929c-b1aecbaf16a4/barbican-api/0.log" Feb 27 17:34:36 crc kubenswrapper[4814]: I0227 17:34:36.806593 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-f8687c5bd-zdnxs_1121d7e8-a956-4a6b-929c-b1aecbaf16a4/barbican-api-log/0.log" Feb 27 17:34:36 crc kubenswrapper[4814]: I0227 17:34:36.897892 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6c5f495fc4-tkkrn_2bbc2d20-3264-487d-8571-71bc22fba348/barbican-keystone-listener-log/0.log" Feb 27 17:34:36 crc kubenswrapper[4814]: I0227 17:34:36.949063 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6c5f495fc4-tkkrn_2bbc2d20-3264-487d-8571-71bc22fba348/barbican-keystone-listener/0.log" Feb 27 17:34:37 crc kubenswrapper[4814]: I0227 17:34:37.079739 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-786585dcfc-sm2x8_b43b2496-ffec-4197-b0df-f03a3fc29ebf/barbican-worker/0.log" Feb 27 17:34:37 crc kubenswrapper[4814]: I0227 17:34:37.111865 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-786585dcfc-sm2x8_b43b2496-ffec-4197-b0df-f03a3fc29ebf/barbican-worker-log/0.log" Feb 27 17:34:37 crc kubenswrapper[4814]: I0227 17:34:37.234178 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-vspjs_556cb206-711d-4002-80cf-1ffe3b8f9643/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:34:37 crc kubenswrapper[4814]: I0227 17:34:37.310053 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_23a14569-9d74-4660-afec-2d91ae3a30a9/ceilometer-central-agent/0.log" Feb 27 17:34:37 crc kubenswrapper[4814]: I0227 17:34:37.403131 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_23a14569-9d74-4660-afec-2d91ae3a30a9/ceilometer-notification-agent/0.log" Feb 27 17:34:37 crc kubenswrapper[4814]: I0227 17:34:37.459584 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_23a14569-9d74-4660-afec-2d91ae3a30a9/proxy-httpd/0.log" Feb 27 17:34:37 crc kubenswrapper[4814]: I0227 17:34:37.530553 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_23a14569-9d74-4660-afec-2d91ae3a30a9/sg-core/0.log" Feb 27 17:34:37 crc kubenswrapper[4814]: I0227 17:34:37.696059 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_41da1428-a337-48a3-9609-35907e0d4955/cinder-api-log/0.log" Feb 27 17:34:37 crc kubenswrapper[4814]: I0227 17:34:37.710213 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_41da1428-a337-48a3-9609-35907e0d4955/cinder-api/0.log" Feb 27 17:34:37 crc kubenswrapper[4814]: I0227 17:34:37.858197 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_bd1ed9cd-1472-46c3-852f-a41a2a316e40/cinder-scheduler/0.log" Feb 27 17:34:37 crc kubenswrapper[4814]: I0227 17:34:37.930163 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_bd1ed9cd-1472-46c3-852f-a41a2a316e40/probe/0.log" Feb 27 17:34:37 crc kubenswrapper[4814]: I0227 17:34:37.996598 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-4dlzw_1c0fc8c4-5559-4dd9-9220-b47c3b537749/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:34:38 crc kubenswrapper[4814]: I0227 17:34:38.135473 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-59k2x_3d27f572-8f1b-4507-a482-f3ce12031e76/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:34:38 crc kubenswrapper[4814]: I0227 17:34:38.240413 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-tww6q_18e28b72-72be-49fd-b568-c3a396e87db0/init/0.log" Feb 27 17:34:38 crc kubenswrapper[4814]: I0227 17:34:38.442636 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-tww6q_18e28b72-72be-49fd-b568-c3a396e87db0/init/0.log" Feb 27 17:34:38 crc kubenswrapper[4814]: I0227 17:34:38.480299 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-wz8bq_daf9ab93-de26-4929-a827-925d1650af5a/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:34:38 crc kubenswrapper[4814]: I0227 17:34:38.536164 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-tww6q_18e28b72-72be-49fd-b568-c3a396e87db0/dnsmasq-dns/0.log" Feb 27 17:34:38 crc kubenswrapper[4814]: I0227 17:34:38.713772 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1e537a87-2e08-4521-bfa5-aeecdc14159d/glance-log/0.log" Feb 27 17:34:38 crc kubenswrapper[4814]: I0227 17:34:38.727310 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1e537a87-2e08-4521-bfa5-aeecdc14159d/glance-httpd/0.log" Feb 27 17:34:38 crc kubenswrapper[4814]: I0227 17:34:38.932115 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_74419238-7f7d-48d0-b138-f56913582843/glance-httpd/0.log" Feb 27 17:34:38 crc kubenswrapper[4814]: I0227 17:34:38.945242 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_74419238-7f7d-48d0-b138-f56913582843/glance-log/0.log" Feb 27 17:34:39 crc kubenswrapper[4814]: I0227 17:34:39.132950 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-65784f76f6-f2pcp_8af960c4-8a04-42d8-83bf-9d03c23ad333/horizon/0.log" Feb 27 17:34:39 crc kubenswrapper[4814]: I0227 17:34:39.247244 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-btx56_d9f3ab2c-0157-4a91-a329-c09648662cce/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:34:39 crc kubenswrapper[4814]: I0227 17:34:39.445886 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-sqtqt_f4d8ff3a-2b21-45ab-8332-5bedf355c09c/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:34:39 crc kubenswrapper[4814]: I0227 17:34:39.534584 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-65784f76f6-f2pcp_8af960c4-8a04-42d8-83bf-9d03c23ad333/horizon-log/0.log" Feb 27 17:34:39 crc kubenswrapper[4814]: I0227 17:34:39.760611 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7877c8f6d6-q56qs_017f253d-4956-4f30-93f1-66ead39f9980/keystone-api/0.log" Feb 27 17:34:39 crc kubenswrapper[4814]: I0227 17:34:39.788226 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29536861-pnnbc_24ebfded-b93e-4720-9747-75b8947d26c1/keystone-cron/0.log" Feb 27 17:34:39 crc kubenswrapper[4814]: I0227 17:34:39.990819 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_971d06e4-1037-4da5-b903-27ddda9e9f00/kube-state-metrics/0.log" Feb 27 17:34:40 crc kubenswrapper[4814]: I0227 17:34:40.041748 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-qqr5n_c96b1d1f-6478-4eea-8413-a0657f19c6e7/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:34:40 crc kubenswrapper[4814]: I0227 17:34:40.439191 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69895c5d4f-4bpj5_9e55471d-9b78-4e06-8195-c73f4ce0d1f6/neutron-api/0.log" Feb 27 17:34:40 crc kubenswrapper[4814]: I0227 17:34:40.476412 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69895c5d4f-4bpj5_9e55471d-9b78-4e06-8195-c73f4ce0d1f6/neutron-httpd/0.log" Feb 27 17:34:40 crc kubenswrapper[4814]: I0227 17:34:40.543716 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-tjs88_cf70174d-7319-418e-8545-83db61a7eb7c/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:34:41 crc kubenswrapper[4814]: I0227 17:34:41.146576 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_1b31cce6-b000-448f-853e-231f707a2be5/nova-api-log/0.log" Feb 27 17:34:41 crc kubenswrapper[4814]: I0227 17:34:41.379218 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_f3152b8d-6ee7-4107-9205-b3535102b3cb/nova-cell0-conductor-conductor/0.log" Feb 27 17:34:41 crc kubenswrapper[4814]: I0227 17:34:41.514589 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_2bac854e-cfa6-4eaa-9272-5ecc27c88d5c/nova-cell1-conductor-conductor/0.log" Feb 27 17:34:41 crc kubenswrapper[4814]: I0227 17:34:41.714341 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_1b31cce6-b000-448f-853e-231f707a2be5/nova-api-api/0.log" Feb 27 17:34:41 crc kubenswrapper[4814]: I0227 17:34:41.798566 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_d4b56f88-0433-4ef4-b4d7-fec94f0f410a/nova-cell1-novncproxy-novncproxy/0.log" Feb 27 17:34:41 crc kubenswrapper[4814]: I0227 17:34:41.828725 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-48n4m_2f3a5797-c53e-4195-8e15-f2ba2016c410/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:34:42 crc kubenswrapper[4814]: I0227 17:34:42.025670 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b7902f26-5f7d-4c34-8166-d1857a631858/nova-metadata-log/0.log" Feb 27 17:34:42 crc kubenswrapper[4814]: I0227 17:34:42.338726 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ea6bae63-9ffa-4ba7-a95f-e686e8d12e87/mysql-bootstrap/0.log" Feb 27 17:34:42 crc kubenswrapper[4814]: I0227 17:34:42.442600 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_aace7eb3-3c5c-4290-8ada-b506bd913841/nova-scheduler-scheduler/0.log" Feb 27 17:34:43 crc kubenswrapper[4814]: I0227 17:34:43.103984 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ea6bae63-9ffa-4ba7-a95f-e686e8d12e87/mysql-bootstrap/0.log" Feb 27 17:34:43 crc kubenswrapper[4814]: I0227 17:34:43.118522 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ea6bae63-9ffa-4ba7-a95f-e686e8d12e87/galera/0.log" Feb 27 17:34:43 crc kubenswrapper[4814]: I0227 17:34:43.313986 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_502232b7-9e95-4215-9dfb-08054b866199/mysql-bootstrap/0.log" Feb 27 17:34:43 crc kubenswrapper[4814]: I0227 17:34:43.516287 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_502232b7-9e95-4215-9dfb-08054b866199/mysql-bootstrap/0.log" Feb 27 17:34:43 crc kubenswrapper[4814]: I0227 17:34:43.531573 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_502232b7-9e95-4215-9dfb-08054b866199/galera/0.log" Feb 27 17:34:43 crc kubenswrapper[4814]: I0227 17:34:43.716108 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_cc735347-66c8-4f38-89e9-31345e59ffee/openstackclient/0.log" Feb 27 17:34:43 crc kubenswrapper[4814]: I0227 17:34:43.734470 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b7902f26-5f7d-4c34-8166-d1857a631858/nova-metadata-metadata/0.log" Feb 27 17:34:43 crc kubenswrapper[4814]: I0227 17:34:43.863606 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-9tkjl_56c74238-2dde-4709-922a-9551da5fe8ae/ovn-controller/0.log" Feb 27 17:34:43 crc kubenswrapper[4814]: I0227 17:34:43.998657 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-hwlpr_b3a71748-678d-498b-8a4b-6bf26679c1cc/openstack-network-exporter/0.log" Feb 27 17:34:44 crc kubenswrapper[4814]: I0227 17:34:44.175462 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kxqxw_62963d98-fb43-4fb1-939a-6f8e6b69cb16/ovsdb-server-init/0.log" Feb 27 17:34:44 crc kubenswrapper[4814]: I0227 17:34:44.321952 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kxqxw_62963d98-fb43-4fb1-939a-6f8e6b69cb16/ovsdb-server-init/0.log" Feb 27 17:34:44 crc kubenswrapper[4814]: I0227 17:34:44.347836 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kxqxw_62963d98-fb43-4fb1-939a-6f8e6b69cb16/ovs-vswitchd/0.log" Feb 27 17:34:44 crc kubenswrapper[4814]: I0227 17:34:44.409502 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kxqxw_62963d98-fb43-4fb1-939a-6f8e6b69cb16/ovsdb-server/0.log" Feb 27 17:34:44 crc kubenswrapper[4814]: I0227 17:34:44.583280 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-58prd_245bd73e-7f47-4c71-9e26-c2b74c8b58e2/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:34:44 crc kubenswrapper[4814]: I0227 17:34:44.677416 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_381df711-c6a7-428a-94d5-8f2f84b51ef6/openstack-network-exporter/0.log" Feb 27 17:34:44 crc kubenswrapper[4814]: I0227 17:34:44.723879 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_381df711-c6a7-428a-94d5-8f2f84b51ef6/ovn-northd/0.log" Feb 27 17:34:45 crc kubenswrapper[4814]: I0227 17:34:45.394968 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_97ad2ba6-a67d-4926-93cd-1caf1c03e493/ovsdbserver-nb/0.log" Feb 27 17:34:45 crc kubenswrapper[4814]: I0227 17:34:45.457071 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_97ad2ba6-a67d-4926-93cd-1caf1c03e493/openstack-network-exporter/0.log" Feb 27 17:34:45 crc kubenswrapper[4814]: I0227 17:34:45.605845 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_bac59d25-01ed-44a7-9750-b654d1c7c631/openstack-network-exporter/0.log" Feb 27 17:34:45 crc kubenswrapper[4814]: I0227 17:34:45.668059 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_bac59d25-01ed-44a7-9750-b654d1c7c631/ovsdbserver-sb/0.log" Feb 27 17:34:45 crc kubenswrapper[4814]: I0227 17:34:45.802160 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8487496c64-cv9xr_13970ebf-9bd1-4292-a552-3f2d6bfa494c/placement-api/0.log" Feb 27 17:34:46 crc kubenswrapper[4814]: I0227 17:34:46.002139 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8487496c64-cv9xr_13970ebf-9bd1-4292-a552-3f2d6bfa494c/placement-log/0.log" Feb 27 17:34:46 crc kubenswrapper[4814]: I0227 17:34:46.027559 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_51ca92ea-753f-4e44-94a8-68e73d165193/setup-container/0.log" Feb 27 17:34:46 crc kubenswrapper[4814]: I0227 17:34:46.257421 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_51ca92ea-753f-4e44-94a8-68e73d165193/setup-container/0.log" Feb 27 17:34:46 crc kubenswrapper[4814]: I0227 17:34:46.257608 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2fb1eae8-a753-43f3-8ed7-759e6e211ace/setup-container/0.log" Feb 27 17:34:46 crc kubenswrapper[4814]: I0227 17:34:46.325556 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_51ca92ea-753f-4e44-94a8-68e73d165193/rabbitmq/0.log" Feb 27 17:34:46 crc kubenswrapper[4814]: I0227 17:34:46.453037 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2fb1eae8-a753-43f3-8ed7-759e6e211ace/setup-container/0.log" Feb 27 17:34:46 crc kubenswrapper[4814]: I0227 17:34:46.467976 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2fb1eae8-a753-43f3-8ed7-759e6e211ace/rabbitmq/0.log" Feb 27 17:34:46 crc kubenswrapper[4814]: I0227 17:34:46.575573 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-4rql9_4cd3437d-9ecc-44b6-b9e7-7aaed24147a3/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:34:46 crc kubenswrapper[4814]: I0227 17:34:46.742237 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-rqbpq_4d3976db-1cd1-4f9b-8975-038c57210a39/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:34:46 crc kubenswrapper[4814]: I0227 17:34:46.805776 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-cz4pg_2bce282e-cd29-40eb-bb24-4ba8503d8a38/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:34:46 crc kubenswrapper[4814]: I0227 17:34:46.960136 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-s5rjk_223a940a-b149-4b9a-a328-411a730cf5fb/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:34:47 crc kubenswrapper[4814]: I0227 17:34:47.111355 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-kz77t_6fddd1b5-7d4c-4986-b192-f06f528c4b64/ssh-known-hosts-edpm-deployment/0.log" Feb 27 17:34:47 crc kubenswrapper[4814]: I0227 17:34:47.353221 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6d4ff56d55-bj24h_6cc6c3ab-803a-45de-a704-2e180b3bd2ce/proxy-server/0.log" Feb 27 17:34:47 crc kubenswrapper[4814]: I0227 17:34:47.412000 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6d4ff56d55-bj24h_6cc6c3ab-803a-45de-a704-2e180b3bd2ce/proxy-httpd/0.log" Feb 27 17:34:47 crc kubenswrapper[4814]: I0227 17:34:47.472302 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-k6ccx_3daf276e-d557-4192-831c-d2aa124fe9bc/swift-ring-rebalance/0.log" Feb 27 17:34:47 crc kubenswrapper[4814]: I0227 17:34:47.642504 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/account-auditor/0.log" Feb 27 17:34:47 crc kubenswrapper[4814]: I0227 17:34:47.676448 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/account-reaper/0.log" Feb 27 17:34:47 crc kubenswrapper[4814]: I0227 17:34:47.737112 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/account-replicator/0.log" Feb 27 17:34:47 crc kubenswrapper[4814]: I0227 17:34:47.833427 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/account-server/0.log" Feb 27 17:34:47 crc kubenswrapper[4814]: I0227 17:34:47.864810 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/container-auditor/0.log" Feb 27 17:34:47 crc kubenswrapper[4814]: I0227 17:34:47.961500 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/container-replicator/0.log" Feb 27 17:34:47 crc kubenswrapper[4814]: I0227 17:34:47.986146 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/container-server/0.log" Feb 27 17:34:48 crc kubenswrapper[4814]: I0227 17:34:48.086720 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/container-updater/0.log" Feb 27 17:34:48 crc kubenswrapper[4814]: I0227 17:34:48.116655 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/object-auditor/0.log" Feb 27 17:34:48 crc kubenswrapper[4814]: I0227 17:34:48.213134 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/object-expirer/0.log" Feb 27 17:34:48 crc kubenswrapper[4814]: I0227 17:34:48.305228 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/object-replicator/0.log" Feb 27 17:34:48 crc kubenswrapper[4814]: I0227 17:34:48.356953 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/object-updater/0.log" Feb 27 17:34:48 crc kubenswrapper[4814]: I0227 17:34:48.370541 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/object-server/0.log" Feb 27 17:34:48 crc kubenswrapper[4814]: I0227 17:34:48.484647 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/rsync/0.log" Feb 27 17:34:48 crc kubenswrapper[4814]: I0227 17:34:48.531278 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9746d8ab-e2c8-419c-8327-6705260b31dd/swift-recon-cron/0.log" Feb 27 17:34:48 crc kubenswrapper[4814]: I0227 17:34:48.703459 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-krg55_fee45495-5dbd-4679-b75d-bd5847b4a0fe/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:34:48 crc kubenswrapper[4814]: I0227 17:34:48.779096 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_c11159aa-3255-4574-81a3-182c8a30d970/tempest-tests-tempest-tests-runner/0.log" Feb 27 17:34:48 crc kubenswrapper[4814]: I0227 17:34:48.915557 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_19e7467f-d219-4179-8f27-c3f0ecc2041f/test-operator-logs-container/0.log" Feb 27 17:34:49 crc kubenswrapper[4814]: I0227 17:34:49.053840 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-phrgr_4edcd988-23c8-4b8c-a3cd-a4d1e727ca3d/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 17:35:00 crc kubenswrapper[4814]: I0227 17:35:00.726870 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_414562e2-0c3b-4db9-bc71-78928f882398/memcached/0.log" Feb 27 17:35:20 crc kubenswrapper[4814]: I0227 17:35:20.992773 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g_0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8/util/0.log" Feb 27 17:35:21 crc kubenswrapper[4814]: I0227 17:35:21.237094 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g_0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8/pull/0.log" Feb 27 17:35:21 crc kubenswrapper[4814]: I0227 17:35:21.240360 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g_0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8/util/0.log" Feb 27 17:35:21 crc kubenswrapper[4814]: I0227 17:35:21.248717 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g_0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8/pull/0.log" Feb 27 17:35:21 crc kubenswrapper[4814]: I0227 17:35:21.462184 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g_0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8/pull/0.log" Feb 27 17:35:21 crc kubenswrapper[4814]: I0227 17:35:21.477249 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g_0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8/util/0.log" Feb 27 17:35:21 crc kubenswrapper[4814]: I0227 17:35:21.523710 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c535b8066aec2871cd66f26395837e1b1c8375ba52a4c2e135cb8cc015gfl4g_0a46f0e7-4a7d-43fc-aa4f-aa9c1a6f5fe8/extract/0.log" Feb 27 17:35:21 crc kubenswrapper[4814]: I0227 17:35:21.974842 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-5d87c9d997-pjncz_6ee0f0c6-52ee-4d20-999f-a0a579dcdaef/manager/0.log" Feb 27 17:35:22 crc kubenswrapper[4814]: I0227 17:35:22.376053 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-64db6967f8-9kqgh_49b03138-9b4c-486c-af91-37c16e8c2536/manager/0.log" Feb 27 17:35:22 crc kubenswrapper[4814]: I0227 17:35:22.474030 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-cf99c678f-ccrlj_a3038d2f-e4de-4d0f-8ac0-e5c666077395/manager/0.log" Feb 27 17:35:22 crc kubenswrapper[4814]: I0227 17:35:22.685072 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-78bc7f9bd9-p74f5_ea6ecef0-2796-4c32-9a9a-ee178574d2f9/manager/0.log" Feb 27 17:35:23 crc kubenswrapper[4814]: I0227 17:35:23.395853 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-545456dc4-fp474_2192fb73-bd5d-440f-8c01-2871658164d9/manager/0.log" Feb 27 17:35:23 crc kubenswrapper[4814]: I0227 17:35:23.441092 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-f7fcc58b9-qc48n_2d989d20-0a79-448b-8b73-ad9378119232/manager/0.log" Feb 27 17:35:23 crc kubenswrapper[4814]: I0227 17:35:23.808054 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55ffd4876b-nxrtn_9b633770-3463-40ee-b5fc-48ec5fb4846c/manager/0.log" Feb 27 17:35:24 crc kubenswrapper[4814]: I0227 17:35:24.109928 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-55d77d7b5c-2d2n7_de6e87db-4283-4aab-a97e-23792fd3b480/manager/0.log" Feb 27 17:35:24 crc kubenswrapper[4814]: I0227 17:35:24.196438 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-67d996989d-c5mcj_dfcfffcb-36da-4853-8ed7-ba76caba9d3c/manager/0.log" Feb 27 17:35:24 crc kubenswrapper[4814]: I0227 17:35:24.852207 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-556b8b874-f4dv9_71ebdc25-566e-4b42-8f48-6f3c52e1125c/manager/0.log" Feb 27 17:35:24 crc kubenswrapper[4814]: I0227 17:35:24.888347 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54688575f-q2gkq_50d18356-e250-46b6-bcd3-053f770f8f58/manager/0.log" Feb 27 17:35:25 crc kubenswrapper[4814]: I0227 17:35:25.130409 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5d86c7ddb7-4m2wv_75ee2f25-f59a-4a60-8245-1b638957d234/manager/0.log" Feb 27 17:35:25 crc kubenswrapper[4814]: I0227 17:35:25.179013 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-74b6b5dc96-m5zqv_5ee5efe1-368e-476e-8516-d9b81d9a38a6/manager/0.log" Feb 27 17:35:25 crc kubenswrapper[4814]: I0227 17:35:25.319502 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c6767dc9ct66dr_edf51635-71a3-4c03-8aef-50212a3b8247/manager/0.log" Feb 27 17:35:25 crc kubenswrapper[4814]: I0227 17:35:25.564486 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-7d5f5556b6-b9hqd_43af9a66-30c9-4410-862d-e07c38625ea7/operator/0.log" Feb 27 17:35:25 crc kubenswrapper[4814]: I0227 17:35:25.746756 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-rnx2c_f571f89f-5528-4a9c-9ea4-f1449e67eb56/registry-server/0.log" Feb 27 17:35:25 crc kubenswrapper[4814]: I0227 17:35:25.955518 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-75684d597f-5958h_c6faf190-d834-45e6-8a18-067067c0bef5/manager/0.log" Feb 27 17:35:26 crc kubenswrapper[4814]: I0227 17:35:26.068004 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-648564c9fc-hbdxk_103a95f2-68f7-41b0-9dea-46cde1679842/manager/0.log" Feb 27 17:35:26 crc kubenswrapper[4814]: I0227 17:35:26.356857 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-lwz52_60ad038c-aad7-4309-89cf-17bf1bda00ac/operator/0.log" Feb 27 17:35:26 crc kubenswrapper[4814]: I0227 17:35:26.467595 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9b9ff9f4d-pq6mn_59cb5f17-969c-4c4f-81bf-a5b0e65acace/manager/0.log" Feb 27 17:35:27 crc kubenswrapper[4814]: I0227 17:35:27.121655 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5fdb694969-4s2gh_3f85fa1d-a7c6-49b8-8f85-d02670fa2849/manager/0.log" Feb 27 17:35:27 crc kubenswrapper[4814]: I0227 17:35:27.239373 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-55b5ff4dbb-tck96_30b75b23-8950-40e1-a515-eaf136867a25/manager/0.log" Feb 27 17:35:27 crc kubenswrapper[4814]: I0227 17:35:27.249556 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-655dd5d898-d7f5g_96d3c178-c0f0-4bd5-b3bd-0e572404d30a/manager/0.log" Feb 27 17:35:27 crc kubenswrapper[4814]: I0227 17:35:27.371992 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-bccc79885-hh455_4b02bae2-c5ac-423e-a2f1-b95de1264f47/manager/0.log" Feb 27 17:35:30 crc kubenswrapper[4814]: I0227 17:35:30.987100 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6db6876945-xnhhg_6ede514a-7a77-4cfe-a746-3d7bd2f5d7fd/manager/0.log" Feb 27 17:35:51 crc kubenswrapper[4814]: I0227 17:35:51.296414 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-ws5nw_3dafc1d6-a249-4d30-ad93-f065110076c8/control-plane-machine-set-operator/0.log" Feb 27 17:35:51 crc kubenswrapper[4814]: I0227 17:35:51.321744 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ndc8k_bc099c2d-2785-4509-bd8b-29469341f56d/kube-rbac-proxy/0.log" Feb 27 17:35:51 crc kubenswrapper[4814]: I0227 17:35:51.495242 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ndc8k_bc099c2d-2785-4509-bd8b-29469341f56d/machine-api-operator/0.log" Feb 27 17:36:00 crc kubenswrapper[4814]: I0227 17:36:00.156310 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536896-j24zv"] Feb 27 17:36:00 crc kubenswrapper[4814]: E0227 17:36:00.157276 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d" containerName="container-00" Feb 27 17:36:00 crc kubenswrapper[4814]: I0227 17:36:00.157290 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d" containerName="container-00" Feb 27 17:36:00 crc kubenswrapper[4814]: I0227 17:36:00.157491 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="adf2e1ff-6a6e-46e3-a8ba-4d0f575a748d" containerName="container-00" Feb 27 17:36:00 crc kubenswrapper[4814]: I0227 17:36:00.158228 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536896-j24zv" Feb 27 17:36:00 crc kubenswrapper[4814]: I0227 17:36:00.162010 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:36:00 crc kubenswrapper[4814]: I0227 17:36:00.162064 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:36:00 crc kubenswrapper[4814]: I0227 17:36:00.163460 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:36:00 crc kubenswrapper[4814]: I0227 17:36:00.175888 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536896-j24zv"] Feb 27 17:36:00 crc kubenswrapper[4814]: I0227 17:36:00.339370 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf8ct\" (UniqueName: \"kubernetes.io/projected/84272e0f-8d02-4bc7-a4f9-3703c20fe654-kube-api-access-rf8ct\") pod \"auto-csr-approver-29536896-j24zv\" (UID: \"84272e0f-8d02-4bc7-a4f9-3703c20fe654\") " pod="openshift-infra/auto-csr-approver-29536896-j24zv" Feb 27 17:36:00 crc kubenswrapper[4814]: I0227 17:36:00.442026 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf8ct\" (UniqueName: \"kubernetes.io/projected/84272e0f-8d02-4bc7-a4f9-3703c20fe654-kube-api-access-rf8ct\") pod \"auto-csr-approver-29536896-j24zv\" (UID: \"84272e0f-8d02-4bc7-a4f9-3703c20fe654\") " pod="openshift-infra/auto-csr-approver-29536896-j24zv" Feb 27 17:36:00 crc kubenswrapper[4814]: I0227 17:36:00.473674 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf8ct\" (UniqueName: \"kubernetes.io/projected/84272e0f-8d02-4bc7-a4f9-3703c20fe654-kube-api-access-rf8ct\") pod \"auto-csr-approver-29536896-j24zv\" (UID: \"84272e0f-8d02-4bc7-a4f9-3703c20fe654\") " pod="openshift-infra/auto-csr-approver-29536896-j24zv" Feb 27 17:36:00 crc kubenswrapper[4814]: I0227 17:36:00.483165 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536896-j24zv" Feb 27 17:36:00 crc kubenswrapper[4814]: I0227 17:36:00.986922 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536896-j24zv"] Feb 27 17:36:01 crc kubenswrapper[4814]: I0227 17:36:01.225615 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536896-j24zv" event={"ID":"84272e0f-8d02-4bc7-a4f9-3703c20fe654","Type":"ContainerStarted","Data":"c753f235119187680fa13e91a580355ab053fa08e0d6e254ca37deaa00cd5481"} Feb 27 17:36:03 crc kubenswrapper[4814]: I0227 17:36:03.258581 4814 generic.go:334] "Generic (PLEG): container finished" podID="84272e0f-8d02-4bc7-a4f9-3703c20fe654" containerID="ede12e05e1ce51472d6f33f282cdbe1c37125252828ecbf952795e8239566492" exitCode=0 Feb 27 17:36:03 crc kubenswrapper[4814]: I0227 17:36:03.258648 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536896-j24zv" event={"ID":"84272e0f-8d02-4bc7-a4f9-3703c20fe654","Type":"ContainerDied","Data":"ede12e05e1ce51472d6f33f282cdbe1c37125252828ecbf952795e8239566492"} Feb 27 17:36:04 crc kubenswrapper[4814]: I0227 17:36:04.669240 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536896-j24zv" Feb 27 17:36:04 crc kubenswrapper[4814]: I0227 17:36:04.850974 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf8ct\" (UniqueName: \"kubernetes.io/projected/84272e0f-8d02-4bc7-a4f9-3703c20fe654-kube-api-access-rf8ct\") pod \"84272e0f-8d02-4bc7-a4f9-3703c20fe654\" (UID: \"84272e0f-8d02-4bc7-a4f9-3703c20fe654\") " Feb 27 17:36:04 crc kubenswrapper[4814]: I0227 17:36:04.860836 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84272e0f-8d02-4bc7-a4f9-3703c20fe654-kube-api-access-rf8ct" (OuterVolumeSpecName: "kube-api-access-rf8ct") pod "84272e0f-8d02-4bc7-a4f9-3703c20fe654" (UID: "84272e0f-8d02-4bc7-a4f9-3703c20fe654"). InnerVolumeSpecName "kube-api-access-rf8ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:36:04 crc kubenswrapper[4814]: I0227 17:36:04.954080 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf8ct\" (UniqueName: \"kubernetes.io/projected/84272e0f-8d02-4bc7-a4f9-3703c20fe654-kube-api-access-rf8ct\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:05 crc kubenswrapper[4814]: I0227 17:36:05.283462 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536896-j24zv" event={"ID":"84272e0f-8d02-4bc7-a4f9-3703c20fe654","Type":"ContainerDied","Data":"c753f235119187680fa13e91a580355ab053fa08e0d6e254ca37deaa00cd5481"} Feb 27 17:36:05 crc kubenswrapper[4814]: I0227 17:36:05.283521 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c753f235119187680fa13e91a580355ab053fa08e0d6e254ca37deaa00cd5481" Feb 27 17:36:05 crc kubenswrapper[4814]: I0227 17:36:05.283597 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536896-j24zv" Feb 27 17:36:05 crc kubenswrapper[4814]: I0227 17:36:05.752550 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536890-r9hrf"] Feb 27 17:36:05 crc kubenswrapper[4814]: I0227 17:36:05.763121 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536890-r9hrf"] Feb 27 17:36:06 crc kubenswrapper[4814]: I0227 17:36:06.511224 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90efa58d-0276-4e00-af07-2a098c5b06a9" path="/var/lib/kubelet/pods/90efa58d-0276-4e00-af07-2a098c5b06a9/volumes" Feb 27 17:36:07 crc kubenswrapper[4814]: I0227 17:36:07.364554 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-wnnxz_9043722e-d6cc-495d-bcf6-ad5b95545ca6/cert-manager-controller/0.log" Feb 27 17:36:07 crc kubenswrapper[4814]: I0227 17:36:07.581650 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-lpq28_b97362bd-6925-4d89-aba9-a8ad9a9cb1bb/cert-manager-webhook/0.log" Feb 27 17:36:07 crc kubenswrapper[4814]: I0227 17:36:07.615040 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-77gdd_c6ddbede-41e8-4a88-8348-a81b242a9b85/cert-manager-cainjector/0.log" Feb 27 17:36:09 crc kubenswrapper[4814]: I0227 17:36:09.354827 4814 scope.go:117] "RemoveContainer" containerID="26dd361bfa4b030b453619c08d34ebda7923226905b5b66c0ec43566e76cd32e" Feb 27 17:36:22 crc kubenswrapper[4814]: I0227 17:36:22.902929 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:36:22 crc kubenswrapper[4814]: I0227 17:36:22.903611 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:36:23 crc kubenswrapper[4814]: I0227 17:36:23.569846 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5dcbbd79cf-5hqsn_ec633892-636c-460c-8c7a-1b594ba270b5/nmstate-console-plugin/0.log" Feb 27 17:36:23 crc kubenswrapper[4814]: I0227 17:36:23.716217 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-6j6pw_694e9d8b-01d0-444c-bf6f-cc4c3cd4f542/nmstate-handler/0.log" Feb 27 17:36:23 crc kubenswrapper[4814]: I0227 17:36:23.730554 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-69594cc75-9xknj_6df39380-a69c-4ef3-94ba-6a5ef31e7f50/nmstate-metrics/0.log" Feb 27 17:36:23 crc kubenswrapper[4814]: I0227 17:36:23.774343 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-69594cc75-9xknj_6df39380-a69c-4ef3-94ba-6a5ef31e7f50/kube-rbac-proxy/0.log" Feb 27 17:36:24 crc kubenswrapper[4814]: I0227 17:36:24.776979 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-75c5dccd6c-r6vt9_154b3bf1-f2f5-4e4b-9110-b2097784f5d8/nmstate-operator/0.log" Feb 27 17:36:24 crc kubenswrapper[4814]: I0227 17:36:24.826887 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-786f45cff4-bbbjp_65270046-4e69-4b5c-b07f-3f401949f32b/nmstate-webhook/0.log" Feb 27 17:36:48 crc kubenswrapper[4814]: I0227 17:36:48.628046 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xxjwn"] Feb 27 17:36:48 crc kubenswrapper[4814]: E0227 17:36:48.629704 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84272e0f-8d02-4bc7-a4f9-3703c20fe654" containerName="oc" Feb 27 17:36:48 crc kubenswrapper[4814]: I0227 17:36:48.629726 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="84272e0f-8d02-4bc7-a4f9-3703c20fe654" containerName="oc" Feb 27 17:36:48 crc kubenswrapper[4814]: I0227 17:36:48.630065 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="84272e0f-8d02-4bc7-a4f9-3703c20fe654" containerName="oc" Feb 27 17:36:48 crc kubenswrapper[4814]: I0227 17:36:48.631979 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:36:48 crc kubenswrapper[4814]: I0227 17:36:48.657695 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xxjwn"] Feb 27 17:36:48 crc kubenswrapper[4814]: I0227 17:36:48.664678 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qswp\" (UniqueName: \"kubernetes.io/projected/db682acd-860e-43df-970b-03a175bd7247-kube-api-access-5qswp\") pod \"community-operators-xxjwn\" (UID: \"db682acd-860e-43df-970b-03a175bd7247\") " pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:36:48 crc kubenswrapper[4814]: I0227 17:36:48.665315 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db682acd-860e-43df-970b-03a175bd7247-catalog-content\") pod \"community-operators-xxjwn\" (UID: \"db682acd-860e-43df-970b-03a175bd7247\") " pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:36:48 crc kubenswrapper[4814]: I0227 17:36:48.665391 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db682acd-860e-43df-970b-03a175bd7247-utilities\") pod \"community-operators-xxjwn\" (UID: \"db682acd-860e-43df-970b-03a175bd7247\") " pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:36:48 crc kubenswrapper[4814]: I0227 17:36:48.767447 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db682acd-860e-43df-970b-03a175bd7247-catalog-content\") pod \"community-operators-xxjwn\" (UID: \"db682acd-860e-43df-970b-03a175bd7247\") " pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:36:48 crc kubenswrapper[4814]: I0227 17:36:48.767499 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db682acd-860e-43df-970b-03a175bd7247-utilities\") pod \"community-operators-xxjwn\" (UID: \"db682acd-860e-43df-970b-03a175bd7247\") " pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:36:48 crc kubenswrapper[4814]: I0227 17:36:48.767559 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qswp\" (UniqueName: \"kubernetes.io/projected/db682acd-860e-43df-970b-03a175bd7247-kube-api-access-5qswp\") pod \"community-operators-xxjwn\" (UID: \"db682acd-860e-43df-970b-03a175bd7247\") " pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:36:48 crc kubenswrapper[4814]: I0227 17:36:48.768004 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db682acd-860e-43df-970b-03a175bd7247-catalog-content\") pod \"community-operators-xxjwn\" (UID: \"db682acd-860e-43df-970b-03a175bd7247\") " pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:36:48 crc kubenswrapper[4814]: I0227 17:36:48.768227 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db682acd-860e-43df-970b-03a175bd7247-utilities\") pod \"community-operators-xxjwn\" (UID: \"db682acd-860e-43df-970b-03a175bd7247\") " pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:36:48 crc kubenswrapper[4814]: I0227 17:36:48.795320 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qswp\" (UniqueName: \"kubernetes.io/projected/db682acd-860e-43df-970b-03a175bd7247-kube-api-access-5qswp\") pod \"community-operators-xxjwn\" (UID: \"db682acd-860e-43df-970b-03a175bd7247\") " pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:36:48 crc kubenswrapper[4814]: I0227 17:36:48.968928 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:36:49 crc kubenswrapper[4814]: I0227 17:36:49.516865 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xxjwn"] Feb 27 17:36:49 crc kubenswrapper[4814]: I0227 17:36:49.848951 4814 generic.go:334] "Generic (PLEG): container finished" podID="db682acd-860e-43df-970b-03a175bd7247" containerID="1e088c65dcc5a994ba64ca464342d2c0a042406f384e830e016057df45e569af" exitCode=0 Feb 27 17:36:49 crc kubenswrapper[4814]: I0227 17:36:49.849052 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxjwn" event={"ID":"db682acd-860e-43df-970b-03a175bd7247","Type":"ContainerDied","Data":"1e088c65dcc5a994ba64ca464342d2c0a042406f384e830e016057df45e569af"} Feb 27 17:36:49 crc kubenswrapper[4814]: I0227 17:36:49.849340 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxjwn" event={"ID":"db682acd-860e-43df-970b-03a175bd7247","Type":"ContainerStarted","Data":"a88dd588a0bdab2f5d2483b75e1808ba9e0515e84b5239856762cee973baa862"} Feb 27 17:36:49 crc kubenswrapper[4814]: I0227 17:36:49.851491 4814 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:36:50 crc kubenswrapper[4814]: I0227 17:36:50.861029 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxjwn" event={"ID":"db682acd-860e-43df-970b-03a175bd7247","Type":"ContainerStarted","Data":"73278e22add7cd837545142b29ca4f23c5bf1f7c447cc85e93f4ab77ea0c1d58"} Feb 27 17:36:51 crc kubenswrapper[4814]: I0227 17:36:51.876927 4814 generic.go:334] "Generic (PLEG): container finished" podID="db682acd-860e-43df-970b-03a175bd7247" containerID="73278e22add7cd837545142b29ca4f23c5bf1f7c447cc85e93f4ab77ea0c1d58" exitCode=0 Feb 27 17:36:51 crc kubenswrapper[4814]: I0227 17:36:51.876995 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxjwn" event={"ID":"db682acd-860e-43df-970b-03a175bd7247","Type":"ContainerDied","Data":"73278e22add7cd837545142b29ca4f23c5bf1f7c447cc85e93f4ab77ea0c1d58"} Feb 27 17:36:52 crc kubenswrapper[4814]: I0227 17:36:52.892233 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxjwn" event={"ID":"db682acd-860e-43df-970b-03a175bd7247","Type":"ContainerStarted","Data":"14a063fbcacb6f9acae4db62874a3296e5e413d218b8a7626376d11b713c0705"} Feb 27 17:36:52 crc kubenswrapper[4814]: I0227 17:36:52.902870 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:36:52 crc kubenswrapper[4814]: I0227 17:36:52.902963 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:36:52 crc kubenswrapper[4814]: I0227 17:36:52.923206 4814 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xxjwn" podStartSLOduration=2.411560292 podStartE2EDuration="4.923180151s" podCreationTimestamp="2026-02-27 17:36:48 +0000 UTC" firstStartedPulling="2026-02-27 17:36:49.851209139 +0000 UTC m=+4422.303833969" lastFinishedPulling="2026-02-27 17:36:52.362828978 +0000 UTC m=+4424.815453828" observedRunningTime="2026-02-27 17:36:52.918280289 +0000 UTC m=+4425.370905129" watchObservedRunningTime="2026-02-27 17:36:52.923180151 +0000 UTC m=+4425.375804991" Feb 27 17:36:57 crc kubenswrapper[4814]: I0227 17:36:57.070085 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-5lqjv_a197efea-0275-49bc-acd5-b043737cd1a8/kube-rbac-proxy/0.log" Feb 27 17:36:57 crc kubenswrapper[4814]: I0227 17:36:57.173784 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-5lqjv_a197efea-0275-49bc-acd5-b043737cd1a8/controller/0.log" Feb 27 17:36:57 crc kubenswrapper[4814]: I0227 17:36:57.258354 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-frr-files/0.log" Feb 27 17:36:57 crc kubenswrapper[4814]: I0227 17:36:57.520182 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-reloader/0.log" Feb 27 17:36:57 crc kubenswrapper[4814]: I0227 17:36:57.544672 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-frr-files/0.log" Feb 27 17:36:57 crc kubenswrapper[4814]: I0227 17:36:57.553857 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-reloader/0.log" Feb 27 17:36:57 crc kubenswrapper[4814]: I0227 17:36:57.582844 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-metrics/0.log" Feb 27 17:36:57 crc kubenswrapper[4814]: I0227 17:36:57.779918 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-frr-files/0.log" Feb 27 17:36:57 crc kubenswrapper[4814]: I0227 17:36:57.791575 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-metrics/0.log" Feb 27 17:36:57 crc kubenswrapper[4814]: I0227 17:36:57.796271 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-metrics/0.log" Feb 27 17:36:57 crc kubenswrapper[4814]: I0227 17:36:57.801834 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-reloader/0.log" Feb 27 17:36:57 crc kubenswrapper[4814]: I0227 17:36:57.965757 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-frr-files/0.log" Feb 27 17:36:57 crc kubenswrapper[4814]: I0227 17:36:57.980747 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-reloader/0.log" Feb 27 17:36:57 crc kubenswrapper[4814]: I0227 17:36:57.989561 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/cp-metrics/0.log" Feb 27 17:36:58 crc kubenswrapper[4814]: I0227 17:36:58.015920 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/controller/0.log" Feb 27 17:36:58 crc kubenswrapper[4814]: I0227 17:36:58.226700 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/frr-metrics/0.log" Feb 27 17:36:58 crc kubenswrapper[4814]: I0227 17:36:58.232914 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/kube-rbac-proxy/0.log" Feb 27 17:36:58 crc kubenswrapper[4814]: I0227 17:36:58.236177 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/kube-rbac-proxy-frr/0.log" Feb 27 17:36:58 crc kubenswrapper[4814]: I0227 17:36:58.462884 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/reloader/0.log" Feb 27 17:36:58 crc kubenswrapper[4814]: I0227 17:36:58.478318 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7f989f654f-lt84m_bbc12f0a-9041-484e-9216-9a618b63257b/frr-k8s-webhook-server/0.log" Feb 27 17:36:58 crc kubenswrapper[4814]: I0227 17:36:58.694481 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-58945cb6fc-nth4f_e4cdb5ef-b877-45a1-bbc9-1c19284b4eb9/manager/0.log" Feb 27 17:36:58 crc kubenswrapper[4814]: I0227 17:36:58.859968 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5d5d4f645d-nwdxv_20a110bc-385c-4cb3-a3e6-ce3568069d69/webhook-server/0.log" Feb 27 17:36:58 crc kubenswrapper[4814]: I0227 17:36:58.971429 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:36:58 crc kubenswrapper[4814]: I0227 17:36:58.971912 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:36:58 crc kubenswrapper[4814]: I0227 17:36:58.988705 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-v59vj_44587c86-594d-410c-939c-f3a1426ae15a/kube-rbac-proxy/0.log" Feb 27 17:36:59 crc kubenswrapper[4814]: I0227 17:36:59.019096 4814 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:36:59 crc kubenswrapper[4814]: I0227 17:36:59.584133 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-v59vj_44587c86-594d-410c-939c-f3a1426ae15a/speaker/0.log" Feb 27 17:36:59 crc kubenswrapper[4814]: I0227 17:36:59.814931 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r8hf_c484c9ee-2da0-4ec8-adf2-5ded9c667133/frr/0.log" Feb 27 17:37:00 crc kubenswrapper[4814]: I0227 17:37:00.041550 4814 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:37:01 crc kubenswrapper[4814]: I0227 17:37:01.571834 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xxjwn"] Feb 27 17:37:03 crc kubenswrapper[4814]: I0227 17:37:03.008574 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xxjwn" podUID="db682acd-860e-43df-970b-03a175bd7247" containerName="registry-server" containerID="cri-o://14a063fbcacb6f9acae4db62874a3296e5e413d218b8a7626376d11b713c0705" gracePeriod=2 Feb 27 17:37:03 crc kubenswrapper[4814]: I0227 17:37:03.486511 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:37:03 crc kubenswrapper[4814]: I0227 17:37:03.631707 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qswp\" (UniqueName: \"kubernetes.io/projected/db682acd-860e-43df-970b-03a175bd7247-kube-api-access-5qswp\") pod \"db682acd-860e-43df-970b-03a175bd7247\" (UID: \"db682acd-860e-43df-970b-03a175bd7247\") " Feb 27 17:37:03 crc kubenswrapper[4814]: I0227 17:37:03.632059 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db682acd-860e-43df-970b-03a175bd7247-utilities\") pod \"db682acd-860e-43df-970b-03a175bd7247\" (UID: \"db682acd-860e-43df-970b-03a175bd7247\") " Feb 27 17:37:03 crc kubenswrapper[4814]: I0227 17:37:03.632108 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db682acd-860e-43df-970b-03a175bd7247-catalog-content\") pod \"db682acd-860e-43df-970b-03a175bd7247\" (UID: \"db682acd-860e-43df-970b-03a175bd7247\") " Feb 27 17:37:03 crc kubenswrapper[4814]: I0227 17:37:03.633279 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db682acd-860e-43df-970b-03a175bd7247-utilities" (OuterVolumeSpecName: "utilities") pod "db682acd-860e-43df-970b-03a175bd7247" (UID: "db682acd-860e-43df-970b-03a175bd7247"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:37:03 crc kubenswrapper[4814]: I0227 17:37:03.643025 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db682acd-860e-43df-970b-03a175bd7247-kube-api-access-5qswp" (OuterVolumeSpecName: "kube-api-access-5qswp") pod "db682acd-860e-43df-970b-03a175bd7247" (UID: "db682acd-860e-43df-970b-03a175bd7247"). InnerVolumeSpecName "kube-api-access-5qswp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:37:03 crc kubenswrapper[4814]: I0227 17:37:03.705975 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db682acd-860e-43df-970b-03a175bd7247-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db682acd-860e-43df-970b-03a175bd7247" (UID: "db682acd-860e-43df-970b-03a175bd7247"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:37:03 crc kubenswrapper[4814]: I0227 17:37:03.735571 4814 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db682acd-860e-43df-970b-03a175bd7247-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:37:03 crc kubenswrapper[4814]: I0227 17:37:03.735630 4814 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db682acd-860e-43df-970b-03a175bd7247-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:37:03 crc kubenswrapper[4814]: I0227 17:37:03.735655 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qswp\" (UniqueName: \"kubernetes.io/projected/db682acd-860e-43df-970b-03a175bd7247-kube-api-access-5qswp\") on node \"crc\" DevicePath \"\"" Feb 27 17:37:04 crc kubenswrapper[4814]: I0227 17:37:04.023958 4814 generic.go:334] "Generic (PLEG): container finished" podID="db682acd-860e-43df-970b-03a175bd7247" containerID="14a063fbcacb6f9acae4db62874a3296e5e413d218b8a7626376d11b713c0705" exitCode=0 Feb 27 17:37:04 crc kubenswrapper[4814]: I0227 17:37:04.024047 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxjwn" event={"ID":"db682acd-860e-43df-970b-03a175bd7247","Type":"ContainerDied","Data":"14a063fbcacb6f9acae4db62874a3296e5e413d218b8a7626376d11b713c0705"} Feb 27 17:37:04 crc kubenswrapper[4814]: I0227 17:37:04.026200 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxjwn" event={"ID":"db682acd-860e-43df-970b-03a175bd7247","Type":"ContainerDied","Data":"a88dd588a0bdab2f5d2483b75e1808ba9e0515e84b5239856762cee973baa862"} Feb 27 17:37:04 crc kubenswrapper[4814]: I0227 17:37:04.024077 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xxjwn" Feb 27 17:37:04 crc kubenswrapper[4814]: I0227 17:37:04.026304 4814 scope.go:117] "RemoveContainer" containerID="14a063fbcacb6f9acae4db62874a3296e5e413d218b8a7626376d11b713c0705" Feb 27 17:37:04 crc kubenswrapper[4814]: I0227 17:37:04.056819 4814 scope.go:117] "RemoveContainer" containerID="73278e22add7cd837545142b29ca4f23c5bf1f7c447cc85e93f4ab77ea0c1d58" Feb 27 17:37:04 crc kubenswrapper[4814]: I0227 17:37:04.085011 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xxjwn"] Feb 27 17:37:04 crc kubenswrapper[4814]: I0227 17:37:04.096962 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xxjwn"] Feb 27 17:37:04 crc kubenswrapper[4814]: I0227 17:37:04.101614 4814 scope.go:117] "RemoveContainer" containerID="1e088c65dcc5a994ba64ca464342d2c0a042406f384e830e016057df45e569af" Feb 27 17:37:04 crc kubenswrapper[4814]: I0227 17:37:04.155074 4814 scope.go:117] "RemoveContainer" containerID="14a063fbcacb6f9acae4db62874a3296e5e413d218b8a7626376d11b713c0705" Feb 27 17:37:04 crc kubenswrapper[4814]: E0227 17:37:04.155780 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14a063fbcacb6f9acae4db62874a3296e5e413d218b8a7626376d11b713c0705\": container with ID starting with 14a063fbcacb6f9acae4db62874a3296e5e413d218b8a7626376d11b713c0705 not found: ID does not exist" containerID="14a063fbcacb6f9acae4db62874a3296e5e413d218b8a7626376d11b713c0705" Feb 27 17:37:04 crc kubenswrapper[4814]: I0227 17:37:04.155856 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14a063fbcacb6f9acae4db62874a3296e5e413d218b8a7626376d11b713c0705"} err="failed to get container status \"14a063fbcacb6f9acae4db62874a3296e5e413d218b8a7626376d11b713c0705\": rpc error: code = NotFound desc = could not find container \"14a063fbcacb6f9acae4db62874a3296e5e413d218b8a7626376d11b713c0705\": container with ID starting with 14a063fbcacb6f9acae4db62874a3296e5e413d218b8a7626376d11b713c0705 not found: ID does not exist" Feb 27 17:37:04 crc kubenswrapper[4814]: I0227 17:37:04.155899 4814 scope.go:117] "RemoveContainer" containerID="73278e22add7cd837545142b29ca4f23c5bf1f7c447cc85e93f4ab77ea0c1d58" Feb 27 17:37:04 crc kubenswrapper[4814]: E0227 17:37:04.156626 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73278e22add7cd837545142b29ca4f23c5bf1f7c447cc85e93f4ab77ea0c1d58\": container with ID starting with 73278e22add7cd837545142b29ca4f23c5bf1f7c447cc85e93f4ab77ea0c1d58 not found: ID does not exist" containerID="73278e22add7cd837545142b29ca4f23c5bf1f7c447cc85e93f4ab77ea0c1d58" Feb 27 17:37:04 crc kubenswrapper[4814]: I0227 17:37:04.156701 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73278e22add7cd837545142b29ca4f23c5bf1f7c447cc85e93f4ab77ea0c1d58"} err="failed to get container status \"73278e22add7cd837545142b29ca4f23c5bf1f7c447cc85e93f4ab77ea0c1d58\": rpc error: code = NotFound desc = could not find container \"73278e22add7cd837545142b29ca4f23c5bf1f7c447cc85e93f4ab77ea0c1d58\": container with ID starting with 73278e22add7cd837545142b29ca4f23c5bf1f7c447cc85e93f4ab77ea0c1d58 not found: ID does not exist" Feb 27 17:37:04 crc kubenswrapper[4814]: I0227 17:37:04.156740 4814 scope.go:117] "RemoveContainer" containerID="1e088c65dcc5a994ba64ca464342d2c0a042406f384e830e016057df45e569af" Feb 27 17:37:04 crc kubenswrapper[4814]: E0227 17:37:04.157224 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e088c65dcc5a994ba64ca464342d2c0a042406f384e830e016057df45e569af\": container with ID starting with 1e088c65dcc5a994ba64ca464342d2c0a042406f384e830e016057df45e569af not found: ID does not exist" containerID="1e088c65dcc5a994ba64ca464342d2c0a042406f384e830e016057df45e569af" Feb 27 17:37:04 crc kubenswrapper[4814]: I0227 17:37:04.157274 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e088c65dcc5a994ba64ca464342d2c0a042406f384e830e016057df45e569af"} err="failed to get container status \"1e088c65dcc5a994ba64ca464342d2c0a042406f384e830e016057df45e569af\": rpc error: code = NotFound desc = could not find container \"1e088c65dcc5a994ba64ca464342d2c0a042406f384e830e016057df45e569af\": container with ID starting with 1e088c65dcc5a994ba64ca464342d2c0a042406f384e830e016057df45e569af not found: ID does not exist" Feb 27 17:37:04 crc kubenswrapper[4814]: I0227 17:37:04.500969 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db682acd-860e-43df-970b-03a175bd7247" path="/var/lib/kubelet/pods/db682acd-860e-43df-970b-03a175bd7247/volumes" Feb 27 17:37:16 crc kubenswrapper[4814]: I0227 17:37:16.492152 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr_4fb39e04-954d-4a82-9d19-ece0f11c5118/util/0.log" Feb 27 17:37:16 crc kubenswrapper[4814]: I0227 17:37:16.720625 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr_4fb39e04-954d-4a82-9d19-ece0f11c5118/pull/0.log" Feb 27 17:37:16 crc kubenswrapper[4814]: I0227 17:37:16.740786 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr_4fb39e04-954d-4a82-9d19-ece0f11c5118/util/0.log" Feb 27 17:37:16 crc kubenswrapper[4814]: I0227 17:37:16.756888 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr_4fb39e04-954d-4a82-9d19-ece0f11c5118/pull/0.log" Feb 27 17:37:16 crc kubenswrapper[4814]: I0227 17:37:16.969278 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr_4fb39e04-954d-4a82-9d19-ece0f11c5118/pull/0.log" Feb 27 17:37:16 crc kubenswrapper[4814]: I0227 17:37:16.973717 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr_4fb39e04-954d-4a82-9d19-ece0f11c5118/util/0.log" Feb 27 17:37:17 crc kubenswrapper[4814]: I0227 17:37:17.011194 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82rt7nr_4fb39e04-954d-4a82-9d19-ece0f11c5118/extract/0.log" Feb 27 17:37:17 crc kubenswrapper[4814]: I0227 17:37:17.170546 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kbbd8_705d8544-aed0-40dd-aac2-185cfc8ecb06/extract-utilities/0.log" Feb 27 17:37:17 crc kubenswrapper[4814]: I0227 17:37:17.586858 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kbbd8_705d8544-aed0-40dd-aac2-185cfc8ecb06/extract-content/0.log" Feb 27 17:37:17 crc kubenswrapper[4814]: I0227 17:37:17.606884 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kbbd8_705d8544-aed0-40dd-aac2-185cfc8ecb06/extract-utilities/0.log" Feb 27 17:37:17 crc kubenswrapper[4814]: I0227 17:37:17.611501 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kbbd8_705d8544-aed0-40dd-aac2-185cfc8ecb06/extract-content/0.log" Feb 27 17:37:17 crc kubenswrapper[4814]: I0227 17:37:17.803746 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kbbd8_705d8544-aed0-40dd-aac2-185cfc8ecb06/extract-utilities/0.log" Feb 27 17:37:17 crc kubenswrapper[4814]: I0227 17:37:17.883139 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kbbd8_705d8544-aed0-40dd-aac2-185cfc8ecb06/extract-content/0.log" Feb 27 17:37:17 crc kubenswrapper[4814]: I0227 17:37:17.966859 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kbbd8_705d8544-aed0-40dd-aac2-185cfc8ecb06/registry-server/0.log" Feb 27 17:37:18 crc kubenswrapper[4814]: I0227 17:37:18.029709 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4nmp_5452c803-ae28-4bd0-a404-5f3d707b14d5/extract-utilities/0.log" Feb 27 17:37:18 crc kubenswrapper[4814]: I0227 17:37:18.202584 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4nmp_5452c803-ae28-4bd0-a404-5f3d707b14d5/extract-content/0.log" Feb 27 17:37:18 crc kubenswrapper[4814]: I0227 17:37:18.207775 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4nmp_5452c803-ae28-4bd0-a404-5f3d707b14d5/extract-utilities/0.log" Feb 27 17:37:18 crc kubenswrapper[4814]: I0227 17:37:18.214272 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4nmp_5452c803-ae28-4bd0-a404-5f3d707b14d5/extract-content/0.log" Feb 27 17:37:19 crc kubenswrapper[4814]: I0227 17:37:19.095157 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4nmp_5452c803-ae28-4bd0-a404-5f3d707b14d5/extract-content/0.log" Feb 27 17:37:19 crc kubenswrapper[4814]: I0227 17:37:19.111587 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4nmp_5452c803-ae28-4bd0-a404-5f3d707b14d5/extract-utilities/0.log" Feb 27 17:37:19 crc kubenswrapper[4814]: I0227 17:37:19.293122 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj_b790a0b4-30f0-40fa-9c06-6c91abe332e0/util/0.log" Feb 27 17:37:19 crc kubenswrapper[4814]: I0227 17:37:19.636166 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj_b790a0b4-30f0-40fa-9c06-6c91abe332e0/pull/0.log" Feb 27 17:37:19 crc kubenswrapper[4814]: I0227 17:37:19.684688 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj_b790a0b4-30f0-40fa-9c06-6c91abe332e0/pull/0.log" Feb 27 17:37:19 crc kubenswrapper[4814]: I0227 17:37:19.696072 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj_b790a0b4-30f0-40fa-9c06-6c91abe332e0/util/0.log" Feb 27 17:37:19 crc kubenswrapper[4814]: I0227 17:37:19.809057 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v4nmp_5452c803-ae28-4bd0-a404-5f3d707b14d5/registry-server/0.log" Feb 27 17:37:19 crc kubenswrapper[4814]: I0227 17:37:19.855397 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj_b790a0b4-30f0-40fa-9c06-6c91abe332e0/util/0.log" Feb 27 17:37:19 crc kubenswrapper[4814]: I0227 17:37:19.902137 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj_b790a0b4-30f0-40fa-9c06-6c91abe332e0/extract/0.log" Feb 27 17:37:19 crc kubenswrapper[4814]: I0227 17:37:19.932206 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4w5mmj_b790a0b4-30f0-40fa-9c06-6c91abe332e0/pull/0.log" Feb 27 17:37:20 crc kubenswrapper[4814]: I0227 17:37:20.053709 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-lvrr5_c5d324fb-1565-4d40-9304-9e2f6adc9604/marketplace-operator/0.log" Feb 27 17:37:20 crc kubenswrapper[4814]: I0227 17:37:20.141134 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fdwvd_01865c94-699f-4fe9-8427-bfe3a2788531/extract-utilities/0.log" Feb 27 17:37:20 crc kubenswrapper[4814]: I0227 17:37:20.317813 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fdwvd_01865c94-699f-4fe9-8427-bfe3a2788531/extract-utilities/0.log" Feb 27 17:37:20 crc kubenswrapper[4814]: I0227 17:37:20.344580 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fdwvd_01865c94-699f-4fe9-8427-bfe3a2788531/extract-content/0.log" Feb 27 17:37:20 crc kubenswrapper[4814]: I0227 17:37:20.345582 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fdwvd_01865c94-699f-4fe9-8427-bfe3a2788531/extract-content/0.log" Feb 27 17:37:20 crc kubenswrapper[4814]: I0227 17:37:20.521891 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fdwvd_01865c94-699f-4fe9-8427-bfe3a2788531/extract-content/0.log" Feb 27 17:37:20 crc kubenswrapper[4814]: I0227 17:37:20.586064 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fdwvd_01865c94-699f-4fe9-8427-bfe3a2788531/extract-utilities/0.log" Feb 27 17:37:20 crc kubenswrapper[4814]: I0227 17:37:20.630783 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fdwvd_01865c94-699f-4fe9-8427-bfe3a2788531/registry-server/0.log" Feb 27 17:37:20 crc kubenswrapper[4814]: I0227 17:37:20.644304 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5rrh_73643f07-8d1a-4321-902c-12a52a2fe4da/extract-utilities/0.log" Feb 27 17:37:21 crc kubenswrapper[4814]: I0227 17:37:21.509082 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5rrh_73643f07-8d1a-4321-902c-12a52a2fe4da/extract-utilities/0.log" Feb 27 17:37:21 crc kubenswrapper[4814]: I0227 17:37:21.530294 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5rrh_73643f07-8d1a-4321-902c-12a52a2fe4da/extract-content/0.log" Feb 27 17:37:21 crc kubenswrapper[4814]: I0227 17:37:21.547222 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5rrh_73643f07-8d1a-4321-902c-12a52a2fe4da/extract-content/0.log" Feb 27 17:37:21 crc kubenswrapper[4814]: I0227 17:37:21.711387 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5rrh_73643f07-8d1a-4321-902c-12a52a2fe4da/extract-utilities/0.log" Feb 27 17:37:21 crc kubenswrapper[4814]: I0227 17:37:21.752281 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5rrh_73643f07-8d1a-4321-902c-12a52a2fe4da/extract-content/0.log" Feb 27 17:37:22 crc kubenswrapper[4814]: I0227 17:37:22.263266 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5rrh_73643f07-8d1a-4321-902c-12a52a2fe4da/registry-server/0.log" Feb 27 17:37:22 crc kubenswrapper[4814]: I0227 17:37:22.902970 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:37:22 crc kubenswrapper[4814]: I0227 17:37:22.903051 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:37:22 crc kubenswrapper[4814]: I0227 17:37:22.903111 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 17:37:22 crc kubenswrapper[4814]: I0227 17:37:22.904395 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1c245446ce2fe408fb6882fd53824f92e0a19fb4974ea6eaf11f89bd2096f9b4"} pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:37:22 crc kubenswrapper[4814]: I0227 17:37:22.904469 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" containerID="cri-o://1c245446ce2fe408fb6882fd53824f92e0a19fb4974ea6eaf11f89bd2096f9b4" gracePeriod=600 Feb 27 17:37:23 crc kubenswrapper[4814]: I0227 17:37:23.234591 4814 generic.go:334] "Generic (PLEG): container finished" podID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerID="1c245446ce2fe408fb6882fd53824f92e0a19fb4974ea6eaf11f89bd2096f9b4" exitCode=0 Feb 27 17:37:23 crc kubenswrapper[4814]: I0227 17:37:23.234704 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerDied","Data":"1c245446ce2fe408fb6882fd53824f92e0a19fb4974ea6eaf11f89bd2096f9b4"} Feb 27 17:37:23 crc kubenswrapper[4814]: I0227 17:37:23.235047 4814 scope.go:117] "RemoveContainer" containerID="e59fcc167d25f3e6f42160c0989590029390acd981fb2ccd9b1a5faa274d2db3" Feb 27 17:37:24 crc kubenswrapper[4814]: I0227 17:37:24.251799 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerStarted","Data":"5741115d02133673246bdeadac428e7bec33d4a3c61c664f83f1a3952bfa31a7"} Feb 27 17:38:00 crc kubenswrapper[4814]: I0227 17:38:00.161451 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536898-kxdn6"] Feb 27 17:38:00 crc kubenswrapper[4814]: E0227 17:38:00.163642 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db682acd-860e-43df-970b-03a175bd7247" containerName="extract-utilities" Feb 27 17:38:00 crc kubenswrapper[4814]: I0227 17:38:00.163661 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="db682acd-860e-43df-970b-03a175bd7247" containerName="extract-utilities" Feb 27 17:38:00 crc kubenswrapper[4814]: E0227 17:38:00.163695 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db682acd-860e-43df-970b-03a175bd7247" containerName="extract-content" Feb 27 17:38:00 crc kubenswrapper[4814]: I0227 17:38:00.163703 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="db682acd-860e-43df-970b-03a175bd7247" containerName="extract-content" Feb 27 17:38:00 crc kubenswrapper[4814]: E0227 17:38:00.163749 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db682acd-860e-43df-970b-03a175bd7247" containerName="registry-server" Feb 27 17:38:00 crc kubenswrapper[4814]: I0227 17:38:00.163756 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="db682acd-860e-43df-970b-03a175bd7247" containerName="registry-server" Feb 27 17:38:00 crc kubenswrapper[4814]: I0227 17:38:00.163971 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="db682acd-860e-43df-970b-03a175bd7247" containerName="registry-server" Feb 27 17:38:00 crc kubenswrapper[4814]: I0227 17:38:00.164752 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536898-kxdn6" Feb 27 17:38:00 crc kubenswrapper[4814]: I0227 17:38:00.167214 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:38:00 crc kubenswrapper[4814]: I0227 17:38:00.169820 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:38:00 crc kubenswrapper[4814]: I0227 17:38:00.169925 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:38:00 crc kubenswrapper[4814]: I0227 17:38:00.174760 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536898-kxdn6"] Feb 27 17:38:00 crc kubenswrapper[4814]: I0227 17:38:00.189635 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hszr5\" (UniqueName: \"kubernetes.io/projected/b35ae6d0-b72a-42c3-a306-447767efaed5-kube-api-access-hszr5\") pod \"auto-csr-approver-29536898-kxdn6\" (UID: \"b35ae6d0-b72a-42c3-a306-447767efaed5\") " pod="openshift-infra/auto-csr-approver-29536898-kxdn6" Feb 27 17:38:00 crc kubenswrapper[4814]: I0227 17:38:00.291832 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hszr5\" (UniqueName: \"kubernetes.io/projected/b35ae6d0-b72a-42c3-a306-447767efaed5-kube-api-access-hszr5\") pod \"auto-csr-approver-29536898-kxdn6\" (UID: \"b35ae6d0-b72a-42c3-a306-447767efaed5\") " pod="openshift-infra/auto-csr-approver-29536898-kxdn6" Feb 27 17:38:00 crc kubenswrapper[4814]: I0227 17:38:00.315649 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hszr5\" (UniqueName: \"kubernetes.io/projected/b35ae6d0-b72a-42c3-a306-447767efaed5-kube-api-access-hszr5\") pod \"auto-csr-approver-29536898-kxdn6\" (UID: \"b35ae6d0-b72a-42c3-a306-447767efaed5\") " pod="openshift-infra/auto-csr-approver-29536898-kxdn6" Feb 27 17:38:00 crc kubenswrapper[4814]: I0227 17:38:00.502616 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536898-kxdn6" Feb 27 17:38:00 crc kubenswrapper[4814]: I0227 17:38:00.981970 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536898-kxdn6"] Feb 27 17:38:00 crc kubenswrapper[4814]: W0227 17:38:00.994513 4814 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb35ae6d0_b72a_42c3_a306_447767efaed5.slice/crio-a9582e1cc56e627335687d56a250b121ba96d6fd49b5d50af61eecfd916eb509 WatchSource:0}: Error finding container a9582e1cc56e627335687d56a250b121ba96d6fd49b5d50af61eecfd916eb509: Status 404 returned error can't find the container with id a9582e1cc56e627335687d56a250b121ba96d6fd49b5d50af61eecfd916eb509 Feb 27 17:38:01 crc kubenswrapper[4814]: I0227 17:38:01.726416 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536898-kxdn6" event={"ID":"b35ae6d0-b72a-42c3-a306-447767efaed5","Type":"ContainerStarted","Data":"a9582e1cc56e627335687d56a250b121ba96d6fd49b5d50af61eecfd916eb509"} Feb 27 17:38:01 crc kubenswrapper[4814]: E0227 17:38:01.961076 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:38:01 crc kubenswrapper[4814]: E0227 17:38:01.961507 4814 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:38:01 crc kubenswrapper[4814]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:38:01 crc kubenswrapper[4814]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hszr5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536898-kxdn6_openshift-infra(b35ae6d0-b72a-42c3-a306-447767efaed5): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:38:01 crc kubenswrapper[4814]: > logger="UnhandledError" Feb 27 17:38:01 crc kubenswrapper[4814]: E0227 17:38:01.962769 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536898-kxdn6" podUID="b35ae6d0-b72a-42c3-a306-447767efaed5" Feb 27 17:38:02 crc kubenswrapper[4814]: E0227 17:38:02.742451 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536898-kxdn6" podUID="b35ae6d0-b72a-42c3-a306-447767efaed5" Feb 27 17:39:16 crc kubenswrapper[4814]: I0227 17:39:16.733192 4814 generic.go:334] "Generic (PLEG): container finished" podID="22630521-470e-44ef-93dd-821e66f8bdfc" containerID="50962e7da8e64b136763b9ceaf45b3e5640ebc722b3f965efdea1113cbd82594" exitCode=0 Feb 27 17:39:16 crc kubenswrapper[4814]: I0227 17:39:16.733319 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jtkl7/must-gather-28ld5" event={"ID":"22630521-470e-44ef-93dd-821e66f8bdfc","Type":"ContainerDied","Data":"50962e7da8e64b136763b9ceaf45b3e5640ebc722b3f965efdea1113cbd82594"} Feb 27 17:39:16 crc kubenswrapper[4814]: I0227 17:39:16.735167 4814 scope.go:117] "RemoveContainer" containerID="50962e7da8e64b136763b9ceaf45b3e5640ebc722b3f965efdea1113cbd82594" Feb 27 17:39:16 crc kubenswrapper[4814]: I0227 17:39:16.919183 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jtkl7_must-gather-28ld5_22630521-470e-44ef-93dd-821e66f8bdfc/gather/0.log" Feb 27 17:39:18 crc kubenswrapper[4814]: E0227 17:39:18.871337 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:39:18 crc kubenswrapper[4814]: E0227 17:39:18.871999 4814 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:39:18 crc kubenswrapper[4814]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:39:18 crc kubenswrapper[4814]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hszr5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536898-kxdn6_openshift-infra(b35ae6d0-b72a-42c3-a306-447767efaed5): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:39:18 crc kubenswrapper[4814]: > logger="UnhandledError" Feb 27 17:39:18 crc kubenswrapper[4814]: E0227 17:39:18.873338 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536898-kxdn6" podUID="b35ae6d0-b72a-42c3-a306-447767efaed5" Feb 27 17:39:29 crc kubenswrapper[4814]: I0227 17:39:29.969873 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jtkl7/must-gather-28ld5"] Feb 27 17:39:29 crc kubenswrapper[4814]: I0227 17:39:29.971042 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-jtkl7/must-gather-28ld5" podUID="22630521-470e-44ef-93dd-821e66f8bdfc" containerName="copy" containerID="cri-o://243a98d073c905306137a55cf4a27a92a89d8e7184e8ddbc48ee54ada521d0fe" gracePeriod=2 Feb 27 17:39:29 crc kubenswrapper[4814]: I0227 17:39:29.980472 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jtkl7/must-gather-28ld5"] Feb 27 17:39:30 crc kubenswrapper[4814]: I0227 17:39:30.691397 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jtkl7_must-gather-28ld5_22630521-470e-44ef-93dd-821e66f8bdfc/copy/0.log" Feb 27 17:39:30 crc kubenswrapper[4814]: I0227 17:39:30.692543 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jtkl7/must-gather-28ld5" Feb 27 17:39:30 crc kubenswrapper[4814]: I0227 17:39:30.818331 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfcxd\" (UniqueName: \"kubernetes.io/projected/22630521-470e-44ef-93dd-821e66f8bdfc-kube-api-access-tfcxd\") pod \"22630521-470e-44ef-93dd-821e66f8bdfc\" (UID: \"22630521-470e-44ef-93dd-821e66f8bdfc\") " Feb 27 17:39:30 crc kubenswrapper[4814]: I0227 17:39:30.818413 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/22630521-470e-44ef-93dd-821e66f8bdfc-must-gather-output\") pod \"22630521-470e-44ef-93dd-821e66f8bdfc\" (UID: \"22630521-470e-44ef-93dd-821e66f8bdfc\") " Feb 27 17:39:30 crc kubenswrapper[4814]: I0227 17:39:30.827676 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22630521-470e-44ef-93dd-821e66f8bdfc-kube-api-access-tfcxd" (OuterVolumeSpecName: "kube-api-access-tfcxd") pod "22630521-470e-44ef-93dd-821e66f8bdfc" (UID: "22630521-470e-44ef-93dd-821e66f8bdfc"). InnerVolumeSpecName "kube-api-access-tfcxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:39:30 crc kubenswrapper[4814]: I0227 17:39:30.896890 4814 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jtkl7_must-gather-28ld5_22630521-470e-44ef-93dd-821e66f8bdfc/copy/0.log" Feb 27 17:39:30 crc kubenswrapper[4814]: I0227 17:39:30.897656 4814 generic.go:334] "Generic (PLEG): container finished" podID="22630521-470e-44ef-93dd-821e66f8bdfc" containerID="243a98d073c905306137a55cf4a27a92a89d8e7184e8ddbc48ee54ada521d0fe" exitCode=143 Feb 27 17:39:30 crc kubenswrapper[4814]: I0227 17:39:30.897752 4814 scope.go:117] "RemoveContainer" containerID="243a98d073c905306137a55cf4a27a92a89d8e7184e8ddbc48ee54ada521d0fe" Feb 27 17:39:30 crc kubenswrapper[4814]: I0227 17:39:30.897777 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jtkl7/must-gather-28ld5" Feb 27 17:39:30 crc kubenswrapper[4814]: I0227 17:39:30.922005 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfcxd\" (UniqueName: \"kubernetes.io/projected/22630521-470e-44ef-93dd-821e66f8bdfc-kube-api-access-tfcxd\") on node \"crc\" DevicePath \"\"" Feb 27 17:39:30 crc kubenswrapper[4814]: I0227 17:39:30.923957 4814 scope.go:117] "RemoveContainer" containerID="50962e7da8e64b136763b9ceaf45b3e5640ebc722b3f965efdea1113cbd82594" Feb 27 17:39:30 crc kubenswrapper[4814]: I0227 17:39:30.996375 4814 scope.go:117] "RemoveContainer" containerID="243a98d073c905306137a55cf4a27a92a89d8e7184e8ddbc48ee54ada521d0fe" Feb 27 17:39:30 crc kubenswrapper[4814]: E0227 17:39:30.997483 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"243a98d073c905306137a55cf4a27a92a89d8e7184e8ddbc48ee54ada521d0fe\": container with ID starting with 243a98d073c905306137a55cf4a27a92a89d8e7184e8ddbc48ee54ada521d0fe not found: ID does not exist" containerID="243a98d073c905306137a55cf4a27a92a89d8e7184e8ddbc48ee54ada521d0fe" Feb 27 17:39:30 crc kubenswrapper[4814]: I0227 17:39:30.997542 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"243a98d073c905306137a55cf4a27a92a89d8e7184e8ddbc48ee54ada521d0fe"} err="failed to get container status \"243a98d073c905306137a55cf4a27a92a89d8e7184e8ddbc48ee54ada521d0fe\": rpc error: code = NotFound desc = could not find container \"243a98d073c905306137a55cf4a27a92a89d8e7184e8ddbc48ee54ada521d0fe\": container with ID starting with 243a98d073c905306137a55cf4a27a92a89d8e7184e8ddbc48ee54ada521d0fe not found: ID does not exist" Feb 27 17:39:30 crc kubenswrapper[4814]: I0227 17:39:30.997582 4814 scope.go:117] "RemoveContainer" containerID="50962e7da8e64b136763b9ceaf45b3e5640ebc722b3f965efdea1113cbd82594" Feb 27 17:39:30 crc kubenswrapper[4814]: E0227 17:39:30.997981 4814 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50962e7da8e64b136763b9ceaf45b3e5640ebc722b3f965efdea1113cbd82594\": container with ID starting with 50962e7da8e64b136763b9ceaf45b3e5640ebc722b3f965efdea1113cbd82594 not found: ID does not exist" containerID="50962e7da8e64b136763b9ceaf45b3e5640ebc722b3f965efdea1113cbd82594" Feb 27 17:39:30 crc kubenswrapper[4814]: I0227 17:39:30.998007 4814 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50962e7da8e64b136763b9ceaf45b3e5640ebc722b3f965efdea1113cbd82594"} err="failed to get container status \"50962e7da8e64b136763b9ceaf45b3e5640ebc722b3f965efdea1113cbd82594\": rpc error: code = NotFound desc = could not find container \"50962e7da8e64b136763b9ceaf45b3e5640ebc722b3f965efdea1113cbd82594\": container with ID starting with 50962e7da8e64b136763b9ceaf45b3e5640ebc722b3f965efdea1113cbd82594 not found: ID does not exist" Feb 27 17:39:31 crc kubenswrapper[4814]: I0227 17:39:31.040944 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22630521-470e-44ef-93dd-821e66f8bdfc-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "22630521-470e-44ef-93dd-821e66f8bdfc" (UID: "22630521-470e-44ef-93dd-821e66f8bdfc"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:39:31 crc kubenswrapper[4814]: I0227 17:39:31.125905 4814 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/22630521-470e-44ef-93dd-821e66f8bdfc-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 27 17:39:31 crc kubenswrapper[4814]: E0227 17:39:31.491309 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536898-kxdn6" podUID="b35ae6d0-b72a-42c3-a306-447767efaed5" Feb 27 17:39:32 crc kubenswrapper[4814]: I0227 17:39:32.499141 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22630521-470e-44ef-93dd-821e66f8bdfc" path="/var/lib/kubelet/pods/22630521-470e-44ef-93dd-821e66f8bdfc/volumes" Feb 27 17:39:45 crc kubenswrapper[4814]: E0227 17:39:45.394304 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:39:45 crc kubenswrapper[4814]: E0227 17:39:45.395247 4814 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:39:45 crc kubenswrapper[4814]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:39:45 crc kubenswrapper[4814]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hszr5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536898-kxdn6_openshift-infra(b35ae6d0-b72a-42c3-a306-447767efaed5): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:39:45 crc kubenswrapper[4814]: > logger="UnhandledError" Feb 27 17:39:45 crc kubenswrapper[4814]: E0227 17:39:45.396518 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536898-kxdn6" podUID="b35ae6d0-b72a-42c3-a306-447767efaed5" Feb 27 17:39:52 crc kubenswrapper[4814]: I0227 17:39:52.902582 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:39:52 crc kubenswrapper[4814]: I0227 17:39:52.903438 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:39:57 crc kubenswrapper[4814]: E0227 17:39:57.491742 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536898-kxdn6" podUID="b35ae6d0-b72a-42c3-a306-447767efaed5" Feb 27 17:40:00 crc kubenswrapper[4814]: I0227 17:40:00.161061 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536900-zl8d4"] Feb 27 17:40:00 crc kubenswrapper[4814]: E0227 17:40:00.161966 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22630521-470e-44ef-93dd-821e66f8bdfc" containerName="copy" Feb 27 17:40:00 crc kubenswrapper[4814]: I0227 17:40:00.161983 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="22630521-470e-44ef-93dd-821e66f8bdfc" containerName="copy" Feb 27 17:40:00 crc kubenswrapper[4814]: E0227 17:40:00.162020 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22630521-470e-44ef-93dd-821e66f8bdfc" containerName="gather" Feb 27 17:40:00 crc kubenswrapper[4814]: I0227 17:40:00.162028 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="22630521-470e-44ef-93dd-821e66f8bdfc" containerName="gather" Feb 27 17:40:00 crc kubenswrapper[4814]: I0227 17:40:00.162257 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="22630521-470e-44ef-93dd-821e66f8bdfc" containerName="copy" Feb 27 17:40:00 crc kubenswrapper[4814]: I0227 17:40:00.162269 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="22630521-470e-44ef-93dd-821e66f8bdfc" containerName="gather" Feb 27 17:40:00 crc kubenswrapper[4814]: I0227 17:40:00.163032 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536900-zl8d4" Feb 27 17:40:00 crc kubenswrapper[4814]: I0227 17:40:00.201338 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8thbk\" (UniqueName: \"kubernetes.io/projected/a407f9ac-3ce7-473c-aaf4-cad7d4f9494b-kube-api-access-8thbk\") pod \"auto-csr-approver-29536900-zl8d4\" (UID: \"a407f9ac-3ce7-473c-aaf4-cad7d4f9494b\") " pod="openshift-infra/auto-csr-approver-29536900-zl8d4" Feb 27 17:40:00 crc kubenswrapper[4814]: I0227 17:40:00.215908 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536900-zl8d4"] Feb 27 17:40:00 crc kubenswrapper[4814]: I0227 17:40:00.302805 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8thbk\" (UniqueName: \"kubernetes.io/projected/a407f9ac-3ce7-473c-aaf4-cad7d4f9494b-kube-api-access-8thbk\") pod \"auto-csr-approver-29536900-zl8d4\" (UID: \"a407f9ac-3ce7-473c-aaf4-cad7d4f9494b\") " pod="openshift-infra/auto-csr-approver-29536900-zl8d4" Feb 27 17:40:00 crc kubenswrapper[4814]: I0227 17:40:00.330995 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8thbk\" (UniqueName: \"kubernetes.io/projected/a407f9ac-3ce7-473c-aaf4-cad7d4f9494b-kube-api-access-8thbk\") pod \"auto-csr-approver-29536900-zl8d4\" (UID: \"a407f9ac-3ce7-473c-aaf4-cad7d4f9494b\") " pod="openshift-infra/auto-csr-approver-29536900-zl8d4" Feb 27 17:40:00 crc kubenswrapper[4814]: I0227 17:40:00.503893 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536900-zl8d4" Feb 27 17:40:01 crc kubenswrapper[4814]: I0227 17:40:01.026610 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536900-zl8d4"] Feb 27 17:40:01 crc kubenswrapper[4814]: I0227 17:40:01.262811 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536900-zl8d4" event={"ID":"a407f9ac-3ce7-473c-aaf4-cad7d4f9494b","Type":"ContainerStarted","Data":"8c921395cf4f8d0661259468a3bb311ee9f608cc1ae58126e8425d97fddacda8"} Feb 27 17:40:02 crc kubenswrapper[4814]: E0227 17:40:02.124693 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:40:02 crc kubenswrapper[4814]: E0227 17:40:02.125619 4814 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:40:02 crc kubenswrapper[4814]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:40:02 crc kubenswrapper[4814]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8thbk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536900-zl8d4_openshift-infra(a407f9ac-3ce7-473c-aaf4-cad7d4f9494b): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:40:02 crc kubenswrapper[4814]: > logger="UnhandledError" Feb 27 17:40:02 crc kubenswrapper[4814]: E0227 17:40:02.126936 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536900-zl8d4" podUID="a407f9ac-3ce7-473c-aaf4-cad7d4f9494b" Feb 27 17:40:02 crc kubenswrapper[4814]: E0227 17:40:02.280490 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536900-zl8d4" podUID="a407f9ac-3ce7-473c-aaf4-cad7d4f9494b" Feb 27 17:40:11 crc kubenswrapper[4814]: E0227 17:40:11.489904 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536898-kxdn6" podUID="b35ae6d0-b72a-42c3-a306-447767efaed5" Feb 27 17:40:15 crc kubenswrapper[4814]: E0227 17:40:15.406302 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:40:15 crc kubenswrapper[4814]: E0227 17:40:15.407461 4814 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:40:15 crc kubenswrapper[4814]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:40:15 crc kubenswrapper[4814]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8thbk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536900-zl8d4_openshift-infra(a407f9ac-3ce7-473c-aaf4-cad7d4f9494b): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:40:15 crc kubenswrapper[4814]: > logger="UnhandledError" Feb 27 17:40:15 crc kubenswrapper[4814]: E0227 17:40:15.408965 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536900-zl8d4" podUID="a407f9ac-3ce7-473c-aaf4-cad7d4f9494b" Feb 27 17:40:22 crc kubenswrapper[4814]: E0227 17:40:22.491660 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536898-kxdn6" podUID="b35ae6d0-b72a-42c3-a306-447767efaed5" Feb 27 17:40:22 crc kubenswrapper[4814]: I0227 17:40:22.902650 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:40:22 crc kubenswrapper[4814]: I0227 17:40:22.902790 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:40:30 crc kubenswrapper[4814]: E0227 17:40:30.493041 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536900-zl8d4" podUID="a407f9ac-3ce7-473c-aaf4-cad7d4f9494b" Feb 27 17:40:35 crc kubenswrapper[4814]: I0227 17:40:35.731090 4814 generic.go:334] "Generic (PLEG): container finished" podID="b35ae6d0-b72a-42c3-a306-447767efaed5" containerID="cb94e5babba22ec8986eb09edc3800da2e5200dfc709599f126b0ff4b0643ce8" exitCode=0 Feb 27 17:40:35 crc kubenswrapper[4814]: I0227 17:40:35.731170 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536898-kxdn6" event={"ID":"b35ae6d0-b72a-42c3-a306-447767efaed5","Type":"ContainerDied","Data":"cb94e5babba22ec8986eb09edc3800da2e5200dfc709599f126b0ff4b0643ce8"} Feb 27 17:40:37 crc kubenswrapper[4814]: I0227 17:40:37.403673 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536898-kxdn6" Feb 27 17:40:37 crc kubenswrapper[4814]: I0227 17:40:37.526688 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hszr5\" (UniqueName: \"kubernetes.io/projected/b35ae6d0-b72a-42c3-a306-447767efaed5-kube-api-access-hszr5\") pod \"b35ae6d0-b72a-42c3-a306-447767efaed5\" (UID: \"b35ae6d0-b72a-42c3-a306-447767efaed5\") " Feb 27 17:40:37 crc kubenswrapper[4814]: I0227 17:40:37.535356 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b35ae6d0-b72a-42c3-a306-447767efaed5-kube-api-access-hszr5" (OuterVolumeSpecName: "kube-api-access-hszr5") pod "b35ae6d0-b72a-42c3-a306-447767efaed5" (UID: "b35ae6d0-b72a-42c3-a306-447767efaed5"). InnerVolumeSpecName "kube-api-access-hszr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:40:37 crc kubenswrapper[4814]: I0227 17:40:37.630769 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hszr5\" (UniqueName: \"kubernetes.io/projected/b35ae6d0-b72a-42c3-a306-447767efaed5-kube-api-access-hszr5\") on node \"crc\" DevicePath \"\"" Feb 27 17:40:37 crc kubenswrapper[4814]: I0227 17:40:37.763688 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536898-kxdn6" event={"ID":"b35ae6d0-b72a-42c3-a306-447767efaed5","Type":"ContainerDied","Data":"a9582e1cc56e627335687d56a250b121ba96d6fd49b5d50af61eecfd916eb509"} Feb 27 17:40:37 crc kubenswrapper[4814]: I0227 17:40:37.764201 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9582e1cc56e627335687d56a250b121ba96d6fd49b5d50af61eecfd916eb509" Feb 27 17:40:37 crc kubenswrapper[4814]: I0227 17:40:37.763882 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536898-kxdn6" Feb 27 17:40:38 crc kubenswrapper[4814]: I0227 17:40:38.513889 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536892-6ktg6"] Feb 27 17:40:38 crc kubenswrapper[4814]: I0227 17:40:38.523786 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536892-6ktg6"] Feb 27 17:40:40 crc kubenswrapper[4814]: I0227 17:40:40.509405 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e9b3520-7570-4c37-8fec-58b0de0af5a9" path="/var/lib/kubelet/pods/2e9b3520-7570-4c37-8fec-58b0de0af5a9/volumes" Feb 27 17:40:44 crc kubenswrapper[4814]: I0227 17:40:44.860367 4814 generic.go:334] "Generic (PLEG): container finished" podID="a407f9ac-3ce7-473c-aaf4-cad7d4f9494b" containerID="7793ccc1ab61362d1fac34d686dd6632cef277ab3be3c18e8060deb06ab312f8" exitCode=0 Feb 27 17:40:44 crc kubenswrapper[4814]: I0227 17:40:44.860506 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536900-zl8d4" event={"ID":"a407f9ac-3ce7-473c-aaf4-cad7d4f9494b","Type":"ContainerDied","Data":"7793ccc1ab61362d1fac34d686dd6632cef277ab3be3c18e8060deb06ab312f8"} Feb 27 17:40:46 crc kubenswrapper[4814]: I0227 17:40:46.326015 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536900-zl8d4" Feb 27 17:40:46 crc kubenswrapper[4814]: I0227 17:40:46.410335 4814 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8thbk\" (UniqueName: \"kubernetes.io/projected/a407f9ac-3ce7-473c-aaf4-cad7d4f9494b-kube-api-access-8thbk\") pod \"a407f9ac-3ce7-473c-aaf4-cad7d4f9494b\" (UID: \"a407f9ac-3ce7-473c-aaf4-cad7d4f9494b\") " Feb 27 17:40:46 crc kubenswrapper[4814]: I0227 17:40:46.420205 4814 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a407f9ac-3ce7-473c-aaf4-cad7d4f9494b-kube-api-access-8thbk" (OuterVolumeSpecName: "kube-api-access-8thbk") pod "a407f9ac-3ce7-473c-aaf4-cad7d4f9494b" (UID: "a407f9ac-3ce7-473c-aaf4-cad7d4f9494b"). InnerVolumeSpecName "kube-api-access-8thbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:40:46 crc kubenswrapper[4814]: I0227 17:40:46.513580 4814 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8thbk\" (UniqueName: \"kubernetes.io/projected/a407f9ac-3ce7-473c-aaf4-cad7d4f9494b-kube-api-access-8thbk\") on node \"crc\" DevicePath \"\"" Feb 27 17:40:46 crc kubenswrapper[4814]: I0227 17:40:46.893160 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536900-zl8d4" event={"ID":"a407f9ac-3ce7-473c-aaf4-cad7d4f9494b","Type":"ContainerDied","Data":"8c921395cf4f8d0661259468a3bb311ee9f608cc1ae58126e8425d97fddacda8"} Feb 27 17:40:46 crc kubenswrapper[4814]: I0227 17:40:46.893224 4814 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c921395cf4f8d0661259468a3bb311ee9f608cc1ae58126e8425d97fddacda8" Feb 27 17:40:46 crc kubenswrapper[4814]: I0227 17:40:46.893268 4814 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536900-zl8d4" Feb 27 17:40:47 crc kubenswrapper[4814]: I0227 17:40:47.414952 4814 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536894-f45jh"] Feb 27 17:40:47 crc kubenswrapper[4814]: I0227 17:40:47.427823 4814 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536894-f45jh"] Feb 27 17:40:48 crc kubenswrapper[4814]: I0227 17:40:48.504772 4814 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5964bc58-dabe-4e8f-859e-824d177aa798" path="/var/lib/kubelet/pods/5964bc58-dabe-4e8f-859e-824d177aa798/volumes" Feb 27 17:40:52 crc kubenswrapper[4814]: I0227 17:40:52.902909 4814 patch_prober.go:28] interesting pod/machine-config-daemon-nlbk2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:40:52 crc kubenswrapper[4814]: I0227 17:40:52.903656 4814 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:40:52 crc kubenswrapper[4814]: I0227 17:40:52.903752 4814 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" Feb 27 17:40:52 crc kubenswrapper[4814]: I0227 17:40:52.905697 4814 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5741115d02133673246bdeadac428e7bec33d4a3c61c664f83f1a3952bfa31a7"} pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:40:52 crc kubenswrapper[4814]: I0227 17:40:52.905820 4814 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerName="machine-config-daemon" containerID="cri-o://5741115d02133673246bdeadac428e7bec33d4a3c61c664f83f1a3952bfa31a7" gracePeriod=600 Feb 27 17:40:53 crc kubenswrapper[4814]: E0227 17:40:53.058589 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:40:53 crc kubenswrapper[4814]: I0227 17:40:53.993176 4814 generic.go:334] "Generic (PLEG): container finished" podID="be713cf4-11a1-42e7-93c6-ab3f748570a4" containerID="5741115d02133673246bdeadac428e7bec33d4a3c61c664f83f1a3952bfa31a7" exitCode=0 Feb 27 17:40:53 crc kubenswrapper[4814]: I0227 17:40:53.993285 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" event={"ID":"be713cf4-11a1-42e7-93c6-ab3f748570a4","Type":"ContainerDied","Data":"5741115d02133673246bdeadac428e7bec33d4a3c61c664f83f1a3952bfa31a7"} Feb 27 17:40:53 crc kubenswrapper[4814]: I0227 17:40:53.993359 4814 scope.go:117] "RemoveContainer" containerID="1c245446ce2fe408fb6882fd53824f92e0a19fb4974ea6eaf11f89bd2096f9b4" Feb 27 17:40:53 crc kubenswrapper[4814]: I0227 17:40:53.994092 4814 scope.go:117] "RemoveContainer" containerID="5741115d02133673246bdeadac428e7bec33d4a3c61c664f83f1a3952bfa31a7" Feb 27 17:40:53 crc kubenswrapper[4814]: E0227 17:40:53.994803 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:41:05 crc kubenswrapper[4814]: I0227 17:41:05.489692 4814 scope.go:117] "RemoveContainer" containerID="5741115d02133673246bdeadac428e7bec33d4a3c61c664f83f1a3952bfa31a7" Feb 27 17:41:05 crc kubenswrapper[4814]: E0227 17:41:05.491486 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:41:09 crc kubenswrapper[4814]: I0227 17:41:09.753759 4814 scope.go:117] "RemoveContainer" containerID="8107e26799afa00a83f85963b78b1a0cc4fabbd17f38d15594f812c7dc61f66d" Feb 27 17:41:09 crc kubenswrapper[4814]: I0227 17:41:09.834551 4814 scope.go:117] "RemoveContainer" containerID="6ed906f1596af25da928a75c1dcce55b92563606220e928058dc6ee2b9afa179" Feb 27 17:41:17 crc kubenswrapper[4814]: I0227 17:41:17.488072 4814 scope.go:117] "RemoveContainer" containerID="5741115d02133673246bdeadac428e7bec33d4a3c61c664f83f1a3952bfa31a7" Feb 27 17:41:17 crc kubenswrapper[4814]: E0227 17:41:17.489365 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:41:32 crc kubenswrapper[4814]: I0227 17:41:32.488510 4814 scope.go:117] "RemoveContainer" containerID="5741115d02133673246bdeadac428e7bec33d4a3c61c664f83f1a3952bfa31a7" Feb 27 17:41:32 crc kubenswrapper[4814]: E0227 17:41:32.490021 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:41:43 crc kubenswrapper[4814]: I0227 17:41:43.488461 4814 scope.go:117] "RemoveContainer" containerID="5741115d02133673246bdeadac428e7bec33d4a3c61c664f83f1a3952bfa31a7" Feb 27 17:41:43 crc kubenswrapper[4814]: E0227 17:41:43.491594 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:41:58 crc kubenswrapper[4814]: I0227 17:41:58.495603 4814 scope.go:117] "RemoveContainer" containerID="5741115d02133673246bdeadac428e7bec33d4a3c61c664f83f1a3952bfa31a7" Feb 27 17:41:58 crc kubenswrapper[4814]: E0227 17:41:58.496757 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.175418 4814 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536902-qqg46"] Feb 27 17:42:00 crc kubenswrapper[4814]: E0227 17:42:00.176850 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a407f9ac-3ce7-473c-aaf4-cad7d4f9494b" containerName="oc" Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.176875 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="a407f9ac-3ce7-473c-aaf4-cad7d4f9494b" containerName="oc" Feb 27 17:42:00 crc kubenswrapper[4814]: E0227 17:42:00.176891 4814 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b35ae6d0-b72a-42c3-a306-447767efaed5" containerName="oc" Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.176901 4814 state_mem.go:107] "Deleted CPUSet assignment" podUID="b35ae6d0-b72a-42c3-a306-447767efaed5" containerName="oc" Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.177349 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="b35ae6d0-b72a-42c3-a306-447767efaed5" containerName="oc" Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.177372 4814 memory_manager.go:354] "RemoveStaleState removing state" podUID="a407f9ac-3ce7-473c-aaf4-cad7d4f9494b" containerName="oc" Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.178610 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536902-qqg46" Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.184342 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.184688 4814 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-m89s2" Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.185251 4814 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.208802 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536902-qqg46"] Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.267176 4814 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlhmf\" (UniqueName: \"kubernetes.io/projected/b9beeb36-0e66-452d-9660-f6ac93814fc3-kube-api-access-rlhmf\") pod \"auto-csr-approver-29536902-qqg46\" (UID: \"b9beeb36-0e66-452d-9660-f6ac93814fc3\") " pod="openshift-infra/auto-csr-approver-29536902-qqg46" Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.369404 4814 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlhmf\" (UniqueName: \"kubernetes.io/projected/b9beeb36-0e66-452d-9660-f6ac93814fc3-kube-api-access-rlhmf\") pod \"auto-csr-approver-29536902-qqg46\" (UID: \"b9beeb36-0e66-452d-9660-f6ac93814fc3\") " pod="openshift-infra/auto-csr-approver-29536902-qqg46" Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.412129 4814 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlhmf\" (UniqueName: \"kubernetes.io/projected/b9beeb36-0e66-452d-9660-f6ac93814fc3-kube-api-access-rlhmf\") pod \"auto-csr-approver-29536902-qqg46\" (UID: \"b9beeb36-0e66-452d-9660-f6ac93814fc3\") " pod="openshift-infra/auto-csr-approver-29536902-qqg46" Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.501690 4814 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536902-qqg46" Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.808104 4814 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.812292 4814 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536902-qqg46"] Feb 27 17:42:00 crc kubenswrapper[4814]: I0227 17:42:00.883351 4814 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536902-qqg46" event={"ID":"b9beeb36-0e66-452d-9660-f6ac93814fc3","Type":"ContainerStarted","Data":"0a3800f4030fba376f7668b922f019a0fe4ff266b9401fd8ae2b08674d361f61"} Feb 27 17:42:03 crc kubenswrapper[4814]: E0227 17:42:03.118680 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:42:03 crc kubenswrapper[4814]: E0227 17:42:03.119572 4814 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:42:03 crc kubenswrapper[4814]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:42:03 crc kubenswrapper[4814]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rlhmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536902-qqg46_openshift-infra(b9beeb36-0e66-452d-9660-f6ac93814fc3): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:42:03 crc kubenswrapper[4814]: > logger="UnhandledError" Feb 27 17:42:03 crc kubenswrapper[4814]: E0227 17:42:03.120716 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536902-qqg46" podUID="b9beeb36-0e66-452d-9660-f6ac93814fc3" Feb 27 17:42:03 crc kubenswrapper[4814]: E0227 17:42:03.925853 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536902-qqg46" podUID="b9beeb36-0e66-452d-9660-f6ac93814fc3" Feb 27 17:42:09 crc kubenswrapper[4814]: I0227 17:42:09.488137 4814 scope.go:117] "RemoveContainer" containerID="5741115d02133673246bdeadac428e7bec33d4a3c61c664f83f1a3952bfa31a7" Feb 27 17:42:09 crc kubenswrapper[4814]: E0227 17:42:09.573904 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:42:15 crc kubenswrapper[4814]: E0227 17:42:15.587665 4814 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:42:15 crc kubenswrapper[4814]: E0227 17:42:15.588562 4814 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:42:15 crc kubenswrapper[4814]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:42:15 crc kubenswrapper[4814]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rlhmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536902-qqg46_openshift-infra(b9beeb36-0e66-452d-9660-f6ac93814fc3): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:42:15 crc kubenswrapper[4814]: > logger="UnhandledError" Feb 27 17:42:15 crc kubenswrapper[4814]: E0227 17:42:15.589878 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536902-qqg46" podUID="b9beeb36-0e66-452d-9660-f6ac93814fc3" Feb 27 17:42:21 crc kubenswrapper[4814]: I0227 17:42:21.488144 4814 scope.go:117] "RemoveContainer" containerID="5741115d02133673246bdeadac428e7bec33d4a3c61c664f83f1a3952bfa31a7" Feb 27 17:42:21 crc kubenswrapper[4814]: E0227 17:42:21.489615 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" Feb 27 17:42:29 crc kubenswrapper[4814]: E0227 17:42:29.493209 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536902-qqg46" podUID="b9beeb36-0e66-452d-9660-f6ac93814fc3" Feb 27 17:42:32 crc kubenswrapper[4814]: I0227 17:42:32.487942 4814 scope.go:117] "RemoveContainer" containerID="5741115d02133673246bdeadac428e7bec33d4a3c61c664f83f1a3952bfa31a7" Feb 27 17:42:32 crc kubenswrapper[4814]: E0227 17:42:32.488953 4814 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nlbk2_openshift-machine-config-operator(be713cf4-11a1-42e7-93c6-ab3f748570a4)\"" pod="openshift-machine-config-operator/machine-config-daemon-nlbk2" podUID="be713cf4-11a1-42e7-93c6-ab3f748570a4" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515150353633024451 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015150353633017366 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015150342034016502 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015150342034015452 5ustar corecore